diff --git a/.gitattributes b/.gitattributes
index 55e3a15d0b58cf70db47b29f4f36992187827302..45e9d12987334d97ac6c0cbb1888fe97db702feb 100644
--- a/.gitattributes
+++ b/.gitattributes
@@ -63,3 +63,14 @@ runs/test-model/checkpoint-160/tokenizer.json filter=lfs diff=lfs merge=lfs -tex
runs/test-model/checkpoint-240/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/test-model/checkpoint-320/tokenizer.json filter=lfs diff=lfs merge=lfs -text
runs/test-model/checkpoint-80/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/baseline/checkpoint-1080/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/baseline/checkpoint-1440/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/baseline/checkpoint-1800/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/baseline/checkpoint-2160/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/baseline/checkpoint-2520/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/baseline/checkpoint-2880/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/baseline/checkpoint-3240/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/baseline/checkpoint-360/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/baseline/checkpoint-3600/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/baseline/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text
+runs/baseline/checkpoint-720/tokenizer.json filter=lfs diff=lfs merge=lfs -text
diff --git a/runs/baseline/checkpoint-1080/chat_template.jinja b/runs/baseline/checkpoint-1080/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/baseline/checkpoint-1080/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/baseline/checkpoint-1080/config.json b/runs/baseline/checkpoint-1080/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/baseline/checkpoint-1080/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/baseline/checkpoint-1080/generation_config.json b/runs/baseline/checkpoint-1080/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/baseline/checkpoint-1080/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/baseline/checkpoint-1080/model.safetensors b/runs/baseline/checkpoint-1080/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..af8dce58d95b2bf66f36168d7a18ed70fbf1f4df
--- /dev/null
+++ b/runs/baseline/checkpoint-1080/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:cd34eed88b2d67d28b6f780a8191a0103043d184014cd0901149410e227b1a76
+size 583356232
diff --git a/runs/baseline/checkpoint-1080/optimizer.pt b/runs/baseline/checkpoint-1080/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..1cb51be1ecbe79e6e68f3b273a495105475d6df2
--- /dev/null
+++ b/runs/baseline/checkpoint-1080/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:838098c21ee71b236879956ab772f05931eed8de49d4db467033e5966af7ca33
+size 1166825803
diff --git a/runs/baseline/checkpoint-1080/rng_state_0.pth b/runs/baseline/checkpoint-1080/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..ff81a4a04de793f1c690eab304502628a6f01a16
--- /dev/null
+++ b/runs/baseline/checkpoint-1080/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3e64ff3db32365cabac6ffad743c22ab2586a3c7538842f851d54fcc9b00320b
+size 14917
diff --git a/runs/baseline/checkpoint-1080/rng_state_1.pth b/runs/baseline/checkpoint-1080/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..2e0fc04a0aae1c0cb2027f4e52b0d2e25910d9a8
--- /dev/null
+++ b/runs/baseline/checkpoint-1080/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:837b65fefb8e1f24a6981b0d4649ff164abbce25b5c9c0d885ed58d7e0986bae
+size 14917
diff --git a/runs/baseline/checkpoint-1080/scheduler.pt b/runs/baseline/checkpoint-1080/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..8736048f38a59bc77b7ec1746ae478e02d0b7c5c
--- /dev/null
+++ b/runs/baseline/checkpoint-1080/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b0e299effdfaa8cee7008c9e7d1521bed82d9b09e5041046b2ae409f4f1f8149
+size 1465
diff --git a/runs/baseline/checkpoint-1080/tokenizer.json b/runs/baseline/checkpoint-1080/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/baseline/checkpoint-1080/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/baseline/checkpoint-1080/tokenizer_config.json b/runs/baseline/checkpoint-1080/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/baseline/checkpoint-1080/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/baseline/checkpoint-1080/trainer_state.json b/runs/baseline/checkpoint-1080/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..595972e0b0591e4742ea62dd80b78a9ea52fd474
--- /dev/null
+++ b/runs/baseline/checkpoint-1080/trainer_state.json
@@ -0,0 +1,7706 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 15.0,
+ "eval_steps": 500,
+ "global_step": 1080,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.12165174633264542,
+ "learning_rate": 0.0,
+ "loss": 12.027063369750977,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.12109383195638657,
+ "learning_rate": 3e-05,
+ "loss": 12.028635025024414,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.12690064311027527,
+ "learning_rate": 6e-05,
+ "loss": 11.894524574279785,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.1537286937236786,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.672262191772461,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.11778244376182556,
+ "learning_rate": 0.00012,
+ "loss": 11.433671951293945,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.10874935984611511,
+ "learning_rate": 0.00015000000000000001,
+ "loss": 11.270309448242188,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.10641921311616898,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 11.155481338500977,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.1048506498336792,
+ "learning_rate": 0.00021000000000000004,
+ "loss": 11.066675186157227,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.10546834021806717,
+ "learning_rate": 0.00024,
+ "loss": 10.952295303344727,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.10614268481731415,
+ "learning_rate": 0.00027,
+ "loss": 10.811386108398438,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10529010742902756,
+ "learning_rate": 0.00030000000000000003,
+ "loss": 10.657712936401367,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.1051444336771965,
+ "learning_rate": 0.00033,
+ "loss": 10.47591781616211,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10420846194028854,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 10.280141830444336,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10323625057935715,
+ "learning_rate": 0.00039000000000000005,
+ "loss": 10.076366424560547,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10255401581525803,
+ "learning_rate": 0.00042000000000000007,
+ "loss": 9.856939315795898,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10151027888059616,
+ "learning_rate": 0.00045,
+ "loss": 9.632134437561035,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.1006326898932457,
+ "learning_rate": 0.00048,
+ "loss": 9.401637077331543,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.09820562601089478,
+ "learning_rate": 0.00051,
+ "loss": 9.177400588989258,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.0956626906991005,
+ "learning_rate": 0.00054,
+ "loss": 8.95518684387207,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.0922091156244278,
+ "learning_rate": 0.00057,
+ "loss": 8.75066089630127,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.08593932539224625,
+ "learning_rate": 0.0006000000000000001,
+ "loss": 8.557976722717285,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.08014766871929169,
+ "learning_rate": 0.00063,
+ "loss": 8.360604286193848,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.07381244748830795,
+ "learning_rate": 0.00066,
+ "loss": 8.210375785827637,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.06455052644014359,
+ "learning_rate": 0.0006900000000000001,
+ "loss": 8.07168960571289,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.054115474224090576,
+ "learning_rate": 0.0007199999999999999,
+ "loss": 7.960069179534912,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.043575264513492584,
+ "learning_rate": 0.00075,
+ "loss": 7.850685119628906,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.02976679615676403,
+ "learning_rate": 0.0007800000000000001,
+ "loss": 7.824377059936523,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.020552165806293488,
+ "learning_rate": 0.0008100000000000001,
+ "loss": 7.7695417404174805,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.018530743196606636,
+ "learning_rate": 0.0008400000000000001,
+ "loss": 7.782247543334961,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.02716820128262043,
+ "learning_rate": 0.00087,
+ "loss": 7.754945278167725,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.03638409823179245,
+ "learning_rate": 0.0009,
+ "loss": 7.799381256103516,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.041579652577638626,
+ "learning_rate": 0.00093,
+ "loss": 7.776629447937012,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.0304960235953331,
+ "learning_rate": 0.00096,
+ "loss": 7.749439239501953,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.053728822618722916,
+ "learning_rate": 0.00099,
+ "loss": 7.745059490203857,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.036322493106126785,
+ "learning_rate": 0.00102,
+ "loss": 7.752074718475342,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.03232959285378456,
+ "learning_rate": 0.00105,
+ "loss": 7.704257011413574,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.03246184438467026,
+ "learning_rate": 0.00108,
+ "loss": 7.667045593261719,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.036655545234680176,
+ "learning_rate": 0.00111,
+ "loss": 7.623445987701416,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.021042725071310997,
+ "learning_rate": 0.00114,
+ "loss": 7.582821846008301,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.022982975468039513,
+ "learning_rate": 0.00117,
+ "loss": 7.58036994934082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.024781864136457443,
+ "learning_rate": 0.0012000000000000001,
+ "loss": 7.5903000831604,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.03667089343070984,
+ "learning_rate": 0.00123,
+ "loss": 7.588233947753906,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.04083017632365227,
+ "learning_rate": 0.00126,
+ "loss": 7.548343658447266,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.02585385926067829,
+ "learning_rate": 0.00129,
+ "loss": 7.604833602905273,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.018793931230902672,
+ "learning_rate": 0.00132,
+ "loss": 7.56705379486084,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.025538936257362366,
+ "learning_rate": 0.00135,
+ "loss": 7.562107086181641,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.02819318138062954,
+ "learning_rate": 0.0013800000000000002,
+ "loss": 7.517213821411133,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.016239255666732788,
+ "learning_rate": 0.00141,
+ "loss": 7.4780755043029785,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.024470701813697815,
+ "learning_rate": 0.0014399999999999999,
+ "loss": 7.505887985229492,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.015499847941100597,
+ "learning_rate": 0.00147,
+ "loss": 7.484469890594482,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.017785200849175453,
+ "learning_rate": 0.0015,
+ "loss": 7.481822967529297,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.015952805057168007,
+ "learning_rate": 0.0015300000000000001,
+ "loss": 7.484804153442383,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.015771260485053062,
+ "learning_rate": 0.0015600000000000002,
+ "loss": 7.4495439529418945,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.024370530620217323,
+ "learning_rate": 0.00159,
+ "loss": 7.471208572387695,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.018173737451434135,
+ "learning_rate": 0.0016200000000000001,
+ "loss": 7.438447952270508,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.020564094185829163,
+ "learning_rate": 0.0016500000000000002,
+ "loss": 7.437922954559326,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.016154944896697998,
+ "learning_rate": 0.0016800000000000003,
+ "loss": 7.429378032684326,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.016817007213830948,
+ "learning_rate": 0.00171,
+ "loss": 7.406428337097168,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.03460947796702385,
+ "learning_rate": 0.00174,
+ "loss": 7.393331527709961,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.15410488843917847,
+ "learning_rate": 0.0017699999999999999,
+ "loss": 7.645856857299805,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.07443148642778397,
+ "learning_rate": 0.0018,
+ "loss": 7.549508094787598,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.0507572740316391,
+ "learning_rate": 0.00183,
+ "loss": 7.517921447753906,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.09682746231555939,
+ "learning_rate": 0.00186,
+ "loss": 7.543978691101074,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.2571517527103424,
+ "learning_rate": 0.00189,
+ "loss": 7.768949508666992,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.0317784883081913,
+ "learning_rate": 0.00192,
+ "loss": 7.509063720703125,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.0353970006108284,
+ "learning_rate": 0.0019500000000000001,
+ "loss": 7.522322654724121,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.03663994371891022,
+ "learning_rate": 0.00198,
+ "loss": 7.504723072052002,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.025481393560767174,
+ "learning_rate": 0.00201,
+ "loss": 7.452597141265869,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.03099915198981762,
+ "learning_rate": 0.00204,
+ "loss": 7.455441474914551,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.022921698167920113,
+ "learning_rate": 0.00207,
+ "loss": 7.4454827308654785,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.032388731837272644,
+ "learning_rate": 0.0021,
+ "loss": 7.467103958129883,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.022428609430789948,
+ "learning_rate": 0.00213,
+ "loss": 7.467540264129639,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.469719409942627,
+ "eval_runtime": 50.9714,
+ "eval_samples_per_second": 47.909,
+ "eval_steps_per_second": 1.511,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.017480911687016487,
+ "learning_rate": 0.00216,
+ "loss": 7.455810546875,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.016675131395459175,
+ "learning_rate": 0.00219,
+ "loss": 7.474848747253418,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.017450228333473206,
+ "learning_rate": 0.00222,
+ "loss": 7.442611217498779,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.016823675483465195,
+ "learning_rate": 0.0022500000000000003,
+ "loss": 7.454628944396973,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.016717543825507164,
+ "learning_rate": 0.00228,
+ "loss": 7.425317764282227,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.01644429750740528,
+ "learning_rate": 0.00231,
+ "loss": 7.4102864265441895,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.013104400597512722,
+ "learning_rate": 0.00234,
+ "loss": 7.4559173583984375,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.01175684668123722,
+ "learning_rate": 0.00237,
+ "loss": 7.438898086547852,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.013736684806644917,
+ "learning_rate": 0.0024000000000000002,
+ "loss": 7.449005603790283,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01176756713539362,
+ "learning_rate": 0.0024300000000000003,
+ "loss": 7.428421974182129,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.011860918253660202,
+ "learning_rate": 0.00246,
+ "loss": 7.411491394042969,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.010093477554619312,
+ "learning_rate": 0.00249,
+ "loss": 7.39311408996582,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.009453288279473782,
+ "learning_rate": 0.00252,
+ "loss": 7.387656211853027,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.010311692021787167,
+ "learning_rate": 0.00255,
+ "loss": 7.401076316833496,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.010053195990622044,
+ "learning_rate": 0.00258,
+ "loss": 7.39655876159668,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.009758037514984608,
+ "learning_rate": 0.00261,
+ "loss": 7.423122882843018,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008394946344196796,
+ "learning_rate": 0.00264,
+ "loss": 7.383601188659668,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.008457187563180923,
+ "learning_rate": 0.00267,
+ "loss": 7.400934219360352,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.010601334273815155,
+ "learning_rate": 0.0027,
+ "loss": 7.400241851806641,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.009009373374283314,
+ "learning_rate": 0.0027300000000000002,
+ "loss": 7.377348899841309,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.00737508200109005,
+ "learning_rate": 0.0027600000000000003,
+ "loss": 7.385881423950195,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.0075711836107075214,
+ "learning_rate": 0.0027900000000000004,
+ "loss": 7.348570823669434,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.00846901722252369,
+ "learning_rate": 0.00282,
+ "loss": 7.364287853240967,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.007160635199397802,
+ "learning_rate": 0.00285,
+ "loss": 7.3881635665893555,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.00682067358866334,
+ "learning_rate": 0.0028799999999999997,
+ "loss": 7.348663330078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.009108101949095726,
+ "learning_rate": 0.00291,
+ "loss": 7.356302261352539,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.015828927978873253,
+ "learning_rate": 0.00294,
+ "loss": 7.347576141357422,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.009683096781373024,
+ "learning_rate": 0.00297,
+ "loss": 7.323967933654785,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.010833270847797394,
+ "learning_rate": 0.003,
+ "loss": 7.3139495849609375,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.014905400574207306,
+ "learning_rate": 0.002999999401781161,
+ "loss": 7.302030086517334,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.023674853146076202,
+ "learning_rate": 0.002999997607125127,
+ "loss": 7.311811923980713,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.050460800528526306,
+ "learning_rate": 0.002999994616033342,
+ "loss": 7.41972541809082,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.02193174883723259,
+ "learning_rate": 0.002999990428508218,
+ "loss": 7.294468879699707,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.036261629313230515,
+ "learning_rate": 0.0029999850445531283,
+ "loss": 7.295568466186523,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.02734929881989956,
+ "learning_rate": 0.002999978464172409,
+ "loss": 7.309239864349365,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.017884107306599617,
+ "learning_rate": 0.002999970687371363,
+ "loss": 7.2427592277526855,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01700308918952942,
+ "learning_rate": 0.0029999617141562573,
+ "loss": 7.241730213165283,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.0209935512393713,
+ "learning_rate": 0.0029999515445343194,
+ "loss": 7.262085437774658,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.023207105696201324,
+ "learning_rate": 0.0029999401785137437,
+ "loss": 7.25450325012207,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.01907353848218918,
+ "learning_rate": 0.002999927616103687,
+ "loss": 7.23149299621582,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.016149666160345078,
+ "learning_rate": 0.0029999138573142715,
+ "loss": 7.210232734680176,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.017513928934931755,
+ "learning_rate": 0.002999898902156582,
+ "loss": 7.218451976776123,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.015002547763288021,
+ "learning_rate": 0.002999882750642667,
+ "loss": 7.169316291809082,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.014233081601560116,
+ "learning_rate": 0.0029998654027855406,
+ "loss": 7.194825172424316,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.01283260714262724,
+ "learning_rate": 0.0029998468585991783,
+ "loss": 7.15669059753418,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.014416799880564213,
+ "learning_rate": 0.0029998271180985223,
+ "loss": 7.143369674682617,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.014337058179080486,
+ "learning_rate": 0.0029998061812994757,
+ "loss": 7.127734184265137,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.016406210139393806,
+ "learning_rate": 0.002999784048218908,
+ "loss": 7.086504936218262,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.01680421642959118,
+ "learning_rate": 0.002999760718874652,
+ "loss": 7.089420318603516,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.010665548034012318,
+ "learning_rate": 0.0029997361932855016,
+ "loss": 7.102463722229004,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013479137793183327,
+ "learning_rate": 0.002999710471471218,
+ "loss": 7.079098701477051,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.012408486567437649,
+ "learning_rate": 0.0029996835534525246,
+ "loss": 7.0790114402771,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.012035196647047997,
+ "learning_rate": 0.0029996554392511094,
+ "loss": 7.058756351470947,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.00899985246360302,
+ "learning_rate": 0.0029996261288896225,
+ "loss": 7.0739240646362305,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.01072906143963337,
+ "learning_rate": 0.002999595622391679,
+ "loss": 7.085537910461426,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.01047342736274004,
+ "learning_rate": 0.0029995639197818577,
+ "loss": 7.060049057006836,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.02342618815600872,
+ "learning_rate": 0.0029995310210857006,
+ "loss": 7.039482116699219,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.011256859637796879,
+ "learning_rate": 0.0029994969263297134,
+ "loss": 7.040412902832031,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.008407019078731537,
+ "learning_rate": 0.0029994616355413666,
+ "loss": 7.0364274978637695,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.011006148532032967,
+ "learning_rate": 0.0029994251487490916,
+ "loss": 6.991468906402588,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.009103255346417427,
+ "learning_rate": 0.0029993874659822875,
+ "loss": 7.001059532165527,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.007875788025557995,
+ "learning_rate": 0.0029993485872713125,
+ "loss": 6.9974260330200195,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.007872611284255981,
+ "learning_rate": 0.0029993085126474913,
+ "loss": 6.980216979980469,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.010071957483887672,
+ "learning_rate": 0.0029992672421431115,
+ "loss": 6.97799015045166,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.01209618803113699,
+ "learning_rate": 0.0029992247757914237,
+ "loss": 6.97202205657959,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.014884858392179012,
+ "learning_rate": 0.0029991811136266425,
+ "loss": 6.976570129394531,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.014953463338315487,
+ "learning_rate": 0.0029991362556839453,
+ "loss": 6.971050262451172,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.008371063508093357,
+ "learning_rate": 0.002999090201999474,
+ "loss": 6.955556869506836,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.011926881037652493,
+ "learning_rate": 0.002999042952610332,
+ "loss": 6.943947792053223,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.013422893360257149,
+ "learning_rate": 0.002998994507554589,
+ "loss": 6.910022735595703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.009672224521636963,
+ "learning_rate": 0.002998944866871275,
+ "loss": 6.8994855880737305,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.012590581551194191,
+ "learning_rate": 0.0029988940306003853,
+ "loss": 6.88709020614624,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.945146083831787,
+ "eval_runtime": 51.5801,
+ "eval_samples_per_second": 47.344,
+ "eval_steps_per_second": 1.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.016554836183786392,
+ "learning_rate": 0.002998841998782877,
+ "loss": 6.916609764099121,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.011085660196840763,
+ "learning_rate": 0.0029987887714606723,
+ "loss": 6.884910583496094,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.008596168830990791,
+ "learning_rate": 0.0029987343486766547,
+ "loss": 6.900066375732422,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.014487302862107754,
+ "learning_rate": 0.0029986787304746715,
+ "loss": 6.8693342208862305,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.015636997297406197,
+ "learning_rate": 0.002998621916899534,
+ "loss": 6.903670787811279,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.013275462202727795,
+ "learning_rate": 0.0029985639079970153,
+ "loss": 6.889361381530762,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.009044744074344635,
+ "learning_rate": 0.002998504703813852,
+ "loss": 6.8549981117248535,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.011327696964144707,
+ "learning_rate": 0.002998444304397745,
+ "loss": 6.856151103973389,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.01310331467539072,
+ "learning_rate": 0.0029983827097973554,
+ "loss": 6.834944725036621,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.010003753937780857,
+ "learning_rate": 0.00299831992006231,
+ "loss": 6.825613975524902,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.009928115643560886,
+ "learning_rate": 0.0029982559352431967,
+ "loss": 6.828935146331787,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.012302150949835777,
+ "learning_rate": 0.0029981907553915674,
+ "loss": 6.821207046508789,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.020555168390274048,
+ "learning_rate": 0.002998124380559936,
+ "loss": 6.8184895515441895,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.011417638510465622,
+ "learning_rate": 0.0029980568108017797,
+ "loss": 6.788754463195801,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.010772978886961937,
+ "learning_rate": 0.002997988046171538,
+ "loss": 6.7861433029174805,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.012753859162330627,
+ "learning_rate": 0.002997918086724614,
+ "loss": 6.788776397705078,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.011327114887535572,
+ "learning_rate": 0.0029978469325173717,
+ "loss": 6.779160022735596,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01235753484070301,
+ "learning_rate": 0.002997774583607139,
+ "loss": 6.799554824829102,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.01417156495153904,
+ "learning_rate": 0.0029977010400522074,
+ "loss": 6.77957820892334,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.021321553736925125,
+ "learning_rate": 0.002997626301911828,
+ "loss": 6.785632133483887,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.03614472970366478,
+ "learning_rate": 0.002997550369246217,
+ "loss": 6.836414337158203,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03934052586555481,
+ "learning_rate": 0.002997473242116551,
+ "loss": 6.854659557342529,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.016240060329437256,
+ "learning_rate": 0.0029973949205849707,
+ "loss": 6.789690971374512,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.031164905056357384,
+ "learning_rate": 0.0029973154047145782,
+ "loss": 6.8363776206970215,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.013051844201982021,
+ "learning_rate": 0.002997234694569438,
+ "loss": 6.803576469421387,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.022611448541283607,
+ "learning_rate": 0.002997152790214577,
+ "loss": 6.762107849121094,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.017812907695770264,
+ "learning_rate": 0.0029970696917159833,
+ "loss": 6.776218414306641,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.015453332103788853,
+ "learning_rate": 0.0029969853991406088,
+ "loss": 6.742618083953857,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.020630091428756714,
+ "learning_rate": 0.0029968999125563655,
+ "loss": 6.765475749969482,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.014672341756522655,
+ "learning_rate": 0.0029968132320321297,
+ "loss": 6.754009246826172,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.016540220007300377,
+ "learning_rate": 0.0029967253576377376,
+ "loss": 6.734752655029297,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.013495649211108685,
+ "learning_rate": 0.0029966362894439874,
+ "loss": 6.752676010131836,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015936989337205887,
+ "learning_rate": 0.0029965460275226405,
+ "loss": 6.725708961486816,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010431243106722832,
+ "learning_rate": 0.0029964545719464194,
+ "loss": 6.698376655578613,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01669321395456791,
+ "learning_rate": 0.0029963619227890083,
+ "loss": 6.711770534515381,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.013308997265994549,
+ "learning_rate": 0.0029962680801250514,
+ "loss": 6.705286979675293,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011648476123809814,
+ "learning_rate": 0.002996173044030159,
+ "loss": 6.690052509307861,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.03942161798477173,
+ "learning_rate": 0.0029960768145808965,
+ "loss": 6.711885452270508,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.01571274921298027,
+ "learning_rate": 0.002995979391854797,
+ "loss": 6.715847969055176,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.00829714722931385,
+ "learning_rate": 0.0029958807759303515,
+ "loss": 6.685308933258057,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.01374727487564087,
+ "learning_rate": 0.0029957809668870114,
+ "loss": 6.695744514465332,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.010857059620320797,
+ "learning_rate": 0.0029956799648051934,
+ "loss": 6.658294677734375,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.008897246792912483,
+ "learning_rate": 0.002995577769766272,
+ "loss": 6.642439842224121,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015880772843956947,
+ "learning_rate": 0.002995474381852584,
+ "loss": 6.674907684326172,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.021623898297548294,
+ "learning_rate": 0.0029953698011474265,
+ "loss": 6.66236686706543,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.03084627166390419,
+ "learning_rate": 0.0029952640277350596,
+ "loss": 6.666715621948242,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.040561504662036896,
+ "learning_rate": 0.002995157061700702,
+ "loss": 6.649062156677246,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.031126515939831734,
+ "learning_rate": 0.0029950489031305345,
+ "loss": 6.640542030334473,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.00872360821813345,
+ "learning_rate": 0.0029949395521116987,
+ "loss": 6.635221481323242,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.023432321846485138,
+ "learning_rate": 0.0029948290087322966,
+ "loss": 6.642196178436279,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.021096013486385345,
+ "learning_rate": 0.0029947172730813913,
+ "loss": 6.643655300140381,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.014494858682155609,
+ "learning_rate": 0.002994604345249006,
+ "loss": 6.612720966339111,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.05225599184632301,
+ "learning_rate": 0.0029944902253261248,
+ "loss": 6.60659646987915,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.014311269856989384,
+ "learning_rate": 0.0029943749134046915,
+ "loss": 6.54373025894165,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.010777472518384457,
+ "learning_rate": 0.0029942584095776115,
+ "loss": 6.599837303161621,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.012498079799115658,
+ "learning_rate": 0.002994140713938749,
+ "loss": 6.589657783508301,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.015775132924318314,
+ "learning_rate": 0.00299402182658293,
+ "loss": 6.577511787414551,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.017978230491280556,
+ "learning_rate": 0.0029939017476059396,
+ "loss": 6.58717155456543,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.026807745918631554,
+ "learning_rate": 0.0029937804771045236,
+ "loss": 6.578145503997803,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.04853282868862152,
+ "learning_rate": 0.0029936580151763872,
+ "loss": 6.609755516052246,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.1057916209101677,
+ "learning_rate": 0.002993534361920195,
+ "loss": 6.870367050170898,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.04925911873579025,
+ "learning_rate": 0.002993409517435574,
+ "loss": 6.871556282043457,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.05288556218147278,
+ "learning_rate": 0.0029932834818231077,
+ "loss": 6.808315753936768,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.03831710293889046,
+ "learning_rate": 0.0029931562551843405,
+ "loss": 6.827617645263672,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.03665117174386978,
+ "learning_rate": 0.002993027837621778,
+ "loss": 6.767284870147705,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.026216736063361168,
+ "learning_rate": 0.0029928982292388826,
+ "loss": 6.733659744262695,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.024796854704618454,
+ "learning_rate": 0.002992767430140078,
+ "loss": 6.759577751159668,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.02410992607474327,
+ "learning_rate": 0.002992635440430747,
+ "loss": 6.701825141906738,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.019924134016036987,
+ "learning_rate": 0.0029925022602172308,
+ "loss": 6.677121162414551,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.022516867145895958,
+ "learning_rate": 0.00299236788960683,
+ "loss": 6.668622970581055,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.025068556889891624,
+ "learning_rate": 0.002992232328707806,
+ "loss": 6.67718505859375,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.02574208751320839,
+ "learning_rate": 0.002992095577629376,
+ "loss": 6.679594039916992,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.657555103302002,
+ "eval_runtime": 51.0199,
+ "eval_samples_per_second": 47.864,
+ "eval_steps_per_second": 1.509,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.017298126593232155,
+ "learning_rate": 0.0029919576364817184,
+ "loss": 6.620774269104004,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.018840325996279716,
+ "learning_rate": 0.002991818505375971,
+ "loss": 6.613316535949707,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.020813113078475,
+ "learning_rate": 0.002991678184424228,
+ "loss": 6.621555328369141,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.026581479236483574,
+ "learning_rate": 0.0029915366737395437,
+ "loss": 6.613471031188965,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.02084030583500862,
+ "learning_rate": 0.0029913939734359306,
+ "loss": 6.559704780578613,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.01327502354979515,
+ "learning_rate": 0.00299125008362836,
+ "loss": 6.564910888671875,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.015091734007000923,
+ "learning_rate": 0.0029911050044327614,
+ "loss": 6.571262836456299,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.015463982708752155,
+ "learning_rate": 0.002990958735966023,
+ "loss": 6.565750598907471,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.013636826537549496,
+ "learning_rate": 0.0029908112783459894,
+ "loss": 6.57981014251709,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.010613273829221725,
+ "learning_rate": 0.0029906626316914655,
+ "loss": 6.538505554199219,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.01804237812757492,
+ "learning_rate": 0.0029905127961222127,
+ "loss": 6.5282511711120605,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.015131509862840176,
+ "learning_rate": 0.0029903617717589506,
+ "loss": 6.531335830688477,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.010518206283450127,
+ "learning_rate": 0.002990209558723358,
+ "loss": 6.521886825561523,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.011360022239387035,
+ "learning_rate": 0.002990056157138069,
+ "loss": 6.528301239013672,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.011556181125342846,
+ "learning_rate": 0.0029899015671266773,
+ "loss": 6.497917652130127,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.01365791168063879,
+ "learning_rate": 0.0029897457888137325,
+ "loss": 6.52849006652832,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.01091043371707201,
+ "learning_rate": 0.0029895888223247426,
+ "loss": 6.492197036743164,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.008484847843647003,
+ "learning_rate": 0.002989430667786173,
+ "loss": 6.497377395629883,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.00836879201233387,
+ "learning_rate": 0.0029892713253254466,
+ "loss": 6.495189666748047,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.009781497530639172,
+ "learning_rate": 0.0029891107950709406,
+ "loss": 6.454085826873779,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.009603648446500301,
+ "learning_rate": 0.0029889490771519934,
+ "loss": 6.499475479125977,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.007991855964064598,
+ "learning_rate": 0.0029887861716988975,
+ "loss": 6.460785865783691,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.008447187952697277,
+ "learning_rate": 0.0029886220788429025,
+ "loss": 6.504582405090332,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.010945498012006283,
+ "learning_rate": 0.0029884567987162156,
+ "loss": 6.49589729309082,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.011096215806901455,
+ "learning_rate": 0.0029882903314519997,
+ "loss": 6.460657119750977,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.013094739057123661,
+ "learning_rate": 0.0029881226771843747,
+ "loss": 6.462450981140137,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.015423037111759186,
+ "learning_rate": 0.0029879538360484165,
+ "loss": 6.45565128326416,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.0182222630828619,
+ "learning_rate": 0.002987783808180157,
+ "loss": 6.452227592468262,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.01981227844953537,
+ "learning_rate": 0.002987612593716585,
+ "loss": 6.477701663970947,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.026733700186014175,
+ "learning_rate": 0.002987440192795645,
+ "loss": 6.452212333679199,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.03504155948758125,
+ "learning_rate": 0.0029872666055562372,
+ "loss": 6.509665489196777,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.027970125898718834,
+ "learning_rate": 0.002987091832138217,
+ "loss": 6.487089157104492,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.013714591972529888,
+ "learning_rate": 0.0029869158726823974,
+ "loss": 6.469569206237793,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.016472695395350456,
+ "learning_rate": 0.0029867387273305453,
+ "loss": 6.461688995361328,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.018457358703017235,
+ "learning_rate": 0.0029865603962253833,
+ "loss": 6.490985870361328,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.011045100167393684,
+ "learning_rate": 0.0029863808795105896,
+ "loss": 6.403701305389404,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.014126897789537907,
+ "learning_rate": 0.0029862001773307972,
+ "loss": 6.379804611206055,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.014015335589647293,
+ "learning_rate": 0.0029860182898315954,
+ "loss": 6.404377460479736,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.010032658465206623,
+ "learning_rate": 0.002985835217159527,
+ "loss": 6.40291690826416,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.015469838865101337,
+ "learning_rate": 0.0029856509594620912,
+ "loss": 6.4414520263671875,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.011106910184025764,
+ "learning_rate": 0.0029854655168877397,
+ "loss": 6.396444320678711,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.01262320950627327,
+ "learning_rate": 0.0029852788895858816,
+ "loss": 6.385222434997559,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.012783976271748543,
+ "learning_rate": 0.002985091077706878,
+ "loss": 6.3818135261535645,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.01224067248404026,
+ "learning_rate": 0.0029849020814020468,
+ "loss": 6.37886905670166,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.010540086776018143,
+ "learning_rate": 0.0029847119008236576,
+ "loss": 6.37799072265625,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.012479913420975208,
+ "learning_rate": 0.0029845205361249366,
+ "loss": 6.365094184875488,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.010080077685415745,
+ "learning_rate": 0.0029843279874600622,
+ "loss": 6.371852874755859,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.013266298919916153,
+ "learning_rate": 0.002984134254984168,
+ "loss": 6.380039215087891,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.01014771405607462,
+ "learning_rate": 0.002983939338853339,
+ "loss": 6.370337963104248,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.008164579048752785,
+ "learning_rate": 0.0029837432392246185,
+ "loss": 6.331899642944336,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.01020442508161068,
+ "learning_rate": 0.002983545956255998,
+ "loss": 6.35077428817749,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.009898579679429531,
+ "learning_rate": 0.002983347490106426,
+ "loss": 6.371985912322998,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.009668919257819653,
+ "learning_rate": 0.002983147840935802,
+ "loss": 6.356776237487793,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.012560366652905941,
+ "learning_rate": 0.0029829470089049815,
+ "loss": 6.357911109924316,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.018435249105095863,
+ "learning_rate": 0.0029827449941757698,
+ "loss": 6.366821765899658,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.029046855866909027,
+ "learning_rate": 0.002982541796910927,
+ "loss": 6.367595672607422,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.043010637164115906,
+ "learning_rate": 0.002982337417274165,
+ "loss": 6.413588523864746,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.045470818877220154,
+ "learning_rate": 0.0029821318554301497,
+ "loss": 6.435015678405762,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.02148726023733616,
+ "learning_rate": 0.002981925111544498,
+ "loss": 6.330894470214844,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.026808688417077065,
+ "learning_rate": 0.00298171718578378,
+ "loss": 6.384839057922363,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.018495213240385056,
+ "learning_rate": 0.002981508078315517,
+ "loss": 6.36277961730957,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.025393597781658173,
+ "learning_rate": 0.0029812977893081834,
+ "loss": 6.3702239990234375,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020856289193034172,
+ "learning_rate": 0.002981086318931206,
+ "loss": 6.354547500610352,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.017827816307544708,
+ "learning_rate": 0.002980873667354962,
+ "loss": 6.339264869689941,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.02639714442193508,
+ "learning_rate": 0.002980659834750781,
+ "loss": 6.3254194259643555,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.015238096937537193,
+ "learning_rate": 0.002980444821290944,
+ "loss": 6.365842819213867,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.012863729149103165,
+ "learning_rate": 0.002980228627148684,
+ "loss": 6.356776714324951,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.014739899896085262,
+ "learning_rate": 0.0029800112524981845,
+ "loss": 6.332481384277344,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.013582171872258186,
+ "learning_rate": 0.0029797926975145797,
+ "loss": 6.311968803405762,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.012704637832939625,
+ "learning_rate": 0.0029795729623739555,
+ "loss": 6.321654319763184,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.012550560757517815,
+ "learning_rate": 0.0029793520472533495,
+ "loss": 6.298409461975098,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.011170165613293648,
+ "learning_rate": 0.002979129952330748,
+ "loss": 6.294830799102783,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.326757907867432,
+ "eval_runtime": 51.6566,
+ "eval_samples_per_second": 47.274,
+ "eval_steps_per_second": 1.491,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.010947446338832378,
+ "learning_rate": 0.0029789066777850897,
+ "loss": 6.286581039428711,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.010097995400428772,
+ "learning_rate": 0.002978682223796262,
+ "loss": 6.288301467895508,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.01224475260823965,
+ "learning_rate": 0.002978456590545104,
+ "loss": 6.264986991882324,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.009734537452459335,
+ "learning_rate": 0.0029782297782134045,
+ "loss": 6.304243087768555,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.01136956550180912,
+ "learning_rate": 0.0029780017869839013,
+ "loss": 6.244938373565674,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.00935293547809124,
+ "learning_rate": 0.002977772617040284,
+ "loss": 6.251852035522461,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.011118784546852112,
+ "learning_rate": 0.0029775422685671894,
+ "loss": 6.256633758544922,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.010837923735380173,
+ "learning_rate": 0.0029773107417502055,
+ "loss": 6.263411045074463,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.011920305900275707,
+ "learning_rate": 0.0029770780367758703,
+ "loss": 6.239311218261719,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.014117483049631119,
+ "learning_rate": 0.0029768441538316685,
+ "loss": 6.253314018249512,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.022502148523926735,
+ "learning_rate": 0.002976609093106036,
+ "loss": 6.246156692504883,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.03348720073699951,
+ "learning_rate": 0.002976372854788357,
+ "loss": 6.255914688110352,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.03427121415734291,
+ "learning_rate": 0.002976135439068965,
+ "loss": 6.289946556091309,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.02407355234026909,
+ "learning_rate": 0.002975896846139141,
+ "loss": 6.288534164428711,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.017861567437648773,
+ "learning_rate": 0.0029756570761911147,
+ "loss": 6.258049964904785,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.019980227574706078,
+ "learning_rate": 0.0029754161294180646,
+ "loss": 6.250895023345947,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.01443127728998661,
+ "learning_rate": 0.0029751740060141177,
+ "loss": 6.212200164794922,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.014975055120885372,
+ "learning_rate": 0.0029749307061743475,
+ "loss": 6.230568885803223,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01408296450972557,
+ "learning_rate": 0.0029746862300947773,
+ "loss": 6.243950366973877,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.011649327352643013,
+ "learning_rate": 0.0029744405779723766,
+ "loss": 6.197744846343994,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.013392188586294651,
+ "learning_rate": 0.002974193750005063,
+ "loss": 6.207277297973633,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.010902997106313705,
+ "learning_rate": 0.002973945746391701,
+ "loss": 6.166373252868652,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.012915392406284809,
+ "learning_rate": 0.0029736965673321032,
+ "loss": 6.186316013336182,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.010772830806672573,
+ "learning_rate": 0.0029734462130270273,
+ "loss": 6.172578811645508,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.010274911299347878,
+ "learning_rate": 0.002973194683678181,
+ "loss": 6.206851005554199,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.013210494071245193,
+ "learning_rate": 0.0029729419794882164,
+ "loss": 6.21115779876709,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.013372313231229782,
+ "learning_rate": 0.0029726881006607326,
+ "loss": 6.18442440032959,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.008105908520519733,
+ "learning_rate": 0.0029724330474002743,
+ "loss": 6.204354286193848,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.014167108573019505,
+ "learning_rate": 0.0029721768199123346,
+ "loss": 6.180662155151367,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.008339803665876389,
+ "learning_rate": 0.002971919418403351,
+ "loss": 6.150079727172852,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.015939386561512947,
+ "learning_rate": 0.0029716608430807073,
+ "loss": 6.175203323364258,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.010777109302580357,
+ "learning_rate": 0.0029714010941527327,
+ "loss": 6.166106224060059,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.011274735443294048,
+ "learning_rate": 0.002971140171828702,
+ "loss": 6.163111686706543,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013626808300614357,
+ "learning_rate": 0.002970878076318836,
+ "loss": 6.140850067138672,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.016907762736082077,
+ "learning_rate": 0.0029706148078343,
+ "loss": 6.192012786865234,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.018817441537976265,
+ "learning_rate": 0.002970350366587205,
+ "loss": 6.162238121032715,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.02829768694937229,
+ "learning_rate": 0.0029700847527906068,
+ "loss": 6.159066200256348,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.02034052088856697,
+ "learning_rate": 0.002969817966658505,
+ "loss": 6.164158344268799,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.01816369965672493,
+ "learning_rate": 0.0029695500084058447,
+ "loss": 6.132316589355469,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.013685261830687523,
+ "learning_rate": 0.0029692808782485146,
+ "loss": 6.152181625366211,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.009791149757802486,
+ "learning_rate": 0.0029690105764033485,
+ "loss": 6.138544082641602,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.014827069826424122,
+ "learning_rate": 0.002968739103088124,
+ "loss": 6.135766983032227,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.01391249243170023,
+ "learning_rate": 0.0029684664585215618,
+ "loss": 6.146172523498535,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.012729899026453495,
+ "learning_rate": 0.002968192642923326,
+ "loss": 6.122613430023193,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.014175458811223507,
+ "learning_rate": 0.002967917656514026,
+ "loss": 6.12574577331543,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.010958840139210224,
+ "learning_rate": 0.002967641499515213,
+ "loss": 6.111088752746582,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.015380816534161568,
+ "learning_rate": 0.0029673641721493817,
+ "loss": 6.100615501403809,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.025339096784591675,
+ "learning_rate": 0.00296708567463997,
+ "loss": 6.139659881591797,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.03695995733141899,
+ "learning_rate": 0.002966806007211358,
+ "loss": 6.160996437072754,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.0408046580851078,
+ "learning_rate": 0.0029665251700888686,
+ "loss": 6.173980712890625,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.025022024288773537,
+ "learning_rate": 0.0029662431634987674,
+ "loss": 6.145794868469238,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.027254154905676842,
+ "learning_rate": 0.002965959987668262,
+ "loss": 6.16928243637085,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.06191903352737427,
+ "learning_rate": 0.0029656756428255017,
+ "loss": 6.138525009155273,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0348956398665905,
+ "learning_rate": 0.002965390129199579,
+ "loss": 6.185477256774902,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.020859677344560623,
+ "learning_rate": 0.002965103447020526,
+ "loss": 6.148846626281738,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.022810814902186394,
+ "learning_rate": 0.0029648155965193174,
+ "loss": 6.1701178550720215,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.017981817945837975,
+ "learning_rate": 0.00296452657792787,
+ "loss": 6.1323771476745605,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.014905866235494614,
+ "learning_rate": 0.00296423639147904,
+ "loss": 6.123426914215088,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.01463468000292778,
+ "learning_rate": 0.002963945037406626,
+ "loss": 6.146122932434082,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.021616583690047264,
+ "learning_rate": 0.002963652515945366,
+ "loss": 6.093459606170654,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.01550968922674656,
+ "learning_rate": 0.00296335882733094,
+ "loss": 6.114929676055908,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.014597428031265736,
+ "learning_rate": 0.0029630639717999675,
+ "loss": 6.113945960998535,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.018249280750751495,
+ "learning_rate": 0.002962767949590008,
+ "loss": 6.130507469177246,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.014160148799419403,
+ "learning_rate": 0.0029624707609395614,
+ "loss": 6.086946487426758,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.015563832595944405,
+ "learning_rate": 0.002962172406088067,
+ "loss": 6.083332538604736,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.012240015901625156,
+ "learning_rate": 0.002961872885275904,
+ "loss": 6.0822906494140625,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.0141651201993227,
+ "learning_rate": 0.002961572198744391,
+ "loss": 6.049562454223633,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.010004645213484764,
+ "learning_rate": 0.0029612703467357866,
+ "loss": 6.067063808441162,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.011790835298597813,
+ "learning_rate": 0.0029609673294932856,
+ "loss": 6.062193870544434,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.011088060215115547,
+ "learning_rate": 0.002960663147261025,
+ "loss": 6.070950031280518,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.010387620888650417,
+ "learning_rate": 0.002960357800284078,
+ "loss": 6.032423973083496,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.012074458412826061,
+ "learning_rate": 0.002960051288808458,
+ "loss": 6.066788673400879,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.097111701965332,
+ "eval_runtime": 51.8742,
+ "eval_samples_per_second": 47.075,
+ "eval_steps_per_second": 1.484,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.013429952785372734,
+ "learning_rate": 0.0029597436130811153,
+ "loss": 6.022172451019287,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.01764705777168274,
+ "learning_rate": 0.0029594347733499387,
+ "loss": 6.029745101928711,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.02403099089860916,
+ "learning_rate": 0.002959124769863755,
+ "loss": 6.04707145690918,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.029030559584498405,
+ "learning_rate": 0.002958813602872328,
+ "loss": 6.05479097366333,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.029980644583702087,
+ "learning_rate": 0.0029585012726263595,
+ "loss": 6.068657875061035,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.021424276754260063,
+ "learning_rate": 0.0029581877793774885,
+ "loss": 6.062008857727051,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.018076395615935326,
+ "learning_rate": 0.002957873123378291,
+ "loss": 6.023279190063477,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.017688792198896408,
+ "learning_rate": 0.002957557304882279,
+ "loss": 6.014641284942627,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.02313626930117607,
+ "learning_rate": 0.0029572403241439026,
+ "loss": 6.026420593261719,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.04045062139630318,
+ "learning_rate": 0.0029569221814185467,
+ "loss": 6.007199764251709,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.015792209655046463,
+ "learning_rate": 0.0029566028769625334,
+ "loss": 6.017209053039551,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.01990479789674282,
+ "learning_rate": 0.002956282411033121,
+ "loss": 5.982841491699219,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.01191043108701706,
+ "learning_rate": 0.002955960783888503,
+ "loss": 6.0035929679870605,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.01732831820845604,
+ "learning_rate": 0.002955637995787807,
+ "loss": 6.025006294250488,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.032948266714811325,
+ "learning_rate": 0.0029553140469910995,
+ "loss": 6.007736682891846,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.011192727833986282,
+ "learning_rate": 0.00295498893775938,
+ "loss": 6.0022687911987305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014646840281784534,
+ "learning_rate": 0.0029546626683545824,
+ "loss": 5.976479530334473,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01589796133339405,
+ "learning_rate": 0.002954335239039576,
+ "loss": 5.980949878692627,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.015129734762012959,
+ "learning_rate": 0.0029540066500781657,
+ "loss": 6.0073394775390625,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.011140518821775913,
+ "learning_rate": 0.002953676901735088,
+ "loss": 5.976587295532227,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.0109480582177639,
+ "learning_rate": 0.0029533459942760174,
+ "loss": 6.001427173614502,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.013346790336072445,
+ "learning_rate": 0.0029530139279675583,
+ "loss": 5.96919059753418,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.012474581599235535,
+ "learning_rate": 0.002952680703077251,
+ "loss": 5.970426559448242,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01140566635876894,
+ "learning_rate": 0.002952346319873569,
+ "loss": 5.989171504974365,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.0137029392644763,
+ "learning_rate": 0.0029520107786259187,
+ "loss": 6.010115623474121,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015517761930823326,
+ "learning_rate": 0.002951674079604639,
+ "loss": 6.003649711608887,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.02173629403114319,
+ "learning_rate": 0.0029513362230810035,
+ "loss": 5.967074871063232,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02683914266526699,
+ "learning_rate": 0.002950997209327216,
+ "loss": 5.994434833526611,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.025865858420729637,
+ "learning_rate": 0.002950657038616415,
+ "loss": 6.010523796081543,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02544897235929966,
+ "learning_rate": 0.002950315711222668,
+ "loss": 5.994078159332275,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.028426019474864006,
+ "learning_rate": 0.0029499732274209777,
+ "loss": 6.020925521850586,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.026174750179052353,
+ "learning_rate": 0.0029496295874872768,
+ "loss": 5.985616683959961,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.02802138403058052,
+ "learning_rate": 0.00294928479169843,
+ "loss": 5.973659038543701,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.019348938018083572,
+ "learning_rate": 0.002948938840332233,
+ "loss": 5.990137577056885,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.016294041648507118,
+ "learning_rate": 0.0029485917336674115,
+ "loss": 5.962082862854004,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.016167785972356796,
+ "learning_rate": 0.002948243471983625,
+ "loss": 5.9594573974609375,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.014119232073426247,
+ "learning_rate": 0.00294789405556146,
+ "loss": 5.894467353820801,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.013805347494781017,
+ "learning_rate": 0.002947543484682436,
+ "loss": 5.964014530181885,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.014496497809886932,
+ "learning_rate": 0.002947191759629001,
+ "loss": 5.989920616149902,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.014813800342381,
+ "learning_rate": 0.0029468388806845333,
+ "loss": 5.9669413566589355,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.01153622381389141,
+ "learning_rate": 0.0029464848481333423,
+ "loss": 5.918018341064453,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.010525977239012718,
+ "learning_rate": 0.0029461296622606647,
+ "loss": 5.892477989196777,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.011410463601350784,
+ "learning_rate": 0.002945773323352667,
+ "loss": 5.925249099731445,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.010949424467980862,
+ "learning_rate": 0.0029454158316964464,
+ "loss": 5.90645694732666,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.00962804164737463,
+ "learning_rate": 0.002945057187580026,
+ "loss": 5.917054176330566,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.011517546139657497,
+ "learning_rate": 0.0029446973912923597,
+ "loss": 5.919445514678955,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01567666605114937,
+ "learning_rate": 0.0029443364431233285,
+ "loss": 5.884111404418945,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.016781140118837357,
+ "learning_rate": 0.002943974343363742,
+ "loss": 5.9252119064331055,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.02244546078145504,
+ "learning_rate": 0.0029436110923053375,
+ "loss": 5.933143615722656,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.028186792507767677,
+ "learning_rate": 0.00294324669024078,
+ "loss": 5.9079132080078125,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.028585223481059074,
+ "learning_rate": 0.002942881137463661,
+ "loss": 5.925262451171875,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.02781485952436924,
+ "learning_rate": 0.0029425144342685003,
+ "loss": 5.963315010070801,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.021486928686499596,
+ "learning_rate": 0.002942146580950744,
+ "loss": 5.930115699768066,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.016890818253159523,
+ "learning_rate": 0.002941777577806765,
+ "loss": 5.921439170837402,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.0169126745313406,
+ "learning_rate": 0.0029414074251338618,
+ "loss": 5.928220272064209,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.014659176580607891,
+ "learning_rate": 0.002941036123230261,
+ "loss": 5.934546947479248,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.015337883494794369,
+ "learning_rate": 0.002940663672395112,
+ "loss": 5.890145301818848,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01254988368600607,
+ "learning_rate": 0.002940290072928493,
+ "loss": 5.875492095947266,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.014658362604677677,
+ "learning_rate": 0.0029399153251314066,
+ "loss": 5.901272773742676,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.017342250794172287,
+ "learning_rate": 0.002939539429305779,
+ "loss": 5.875853538513184,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.016825608909130096,
+ "learning_rate": 0.0029391623857544643,
+ "loss": 5.941972732543945,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.016716884449124336,
+ "learning_rate": 0.002938784194781238,
+ "loss": 5.912367343902588,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.021707067266106606,
+ "learning_rate": 0.002938404856690803,
+ "loss": 5.895319938659668,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.030519135296344757,
+ "learning_rate": 0.002938024371788785,
+ "loss": 5.897165298461914,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.02572188340127468,
+ "learning_rate": 0.002937642740381733,
+ "loss": 5.879164695739746,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.016286717727780342,
+ "learning_rate": 0.002937259962777122,
+ "loss": 5.841421127319336,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.020656082779169083,
+ "learning_rate": 0.002936876039283347,
+ "loss": 5.9222259521484375,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.015350459143519402,
+ "learning_rate": 0.0029364909702097298,
+ "loss": 5.885847091674805,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01708570495247841,
+ "learning_rate": 0.0029361047558665136,
+ "loss": 5.872044563293457,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.013537376187741756,
+ "learning_rate": 0.0029357173965648634,
+ "loss": 5.866434097290039,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.013208637945353985,
+ "learning_rate": 0.002935328892616868,
+ "loss": 5.8624491691589355,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.012249554507434368,
+ "learning_rate": 0.002934939244335538,
+ "loss": 5.833232879638672,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.90985107421875,
+ "eval_runtime": 51.0696,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.01382362935692072,
+ "learning_rate": 0.0029345484520348062,
+ "loss": 5.835695266723633,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.012947693467140198,
+ "learning_rate": 0.0029341565160295262,
+ "loss": 5.826190948486328,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.013653872534632683,
+ "learning_rate": 0.002933763436635475,
+ "loss": 5.831789016723633,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.027596980333328247,
+ "learning_rate": 0.0029333692141693484,
+ "loss": 5.839963436126709,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01637990213930607,
+ "learning_rate": 0.0029329738489487648,
+ "loss": 5.805248260498047,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.02169182151556015,
+ "learning_rate": 0.0029325773412922627,
+ "loss": 5.838248252868652,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.03518201410770416,
+ "learning_rate": 0.0029321796915193015,
+ "loss": 5.834336280822754,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04983877018094063,
+ "learning_rate": 0.0029317808999502603,
+ "loss": 5.879856109619141,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05085664615035057,
+ "learning_rate": 0.0029313809669064377,
+ "loss": 5.946531295776367,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.03314858302474022,
+ "learning_rate": 0.0029309798927100532,
+ "loss": 5.909746170043945,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.034118980169296265,
+ "learning_rate": 0.0029305776776842453,
+ "loss": 5.903947830200195,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.03401247411966324,
+ "learning_rate": 0.0029301743221530705,
+ "loss": 5.936162948608398,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.043827857822179794,
+ "learning_rate": 0.0029297698264415047,
+ "loss": 5.953949451446533,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.06281330436468124,
+ "learning_rate": 0.002929364190875444,
+ "loss": 5.988225936889648,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.07216024398803711,
+ "learning_rate": 0.0029289574157817014,
+ "loss": 6.187741279602051,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.055205658078193665,
+ "learning_rate": 0.002928549501488008,
+ "loss": 6.177831649780273,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.04834286496043205,
+ "learning_rate": 0.0029281404483230133,
+ "loss": 6.209604263305664,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.0447777584195137,
+ "learning_rate": 0.0029277302566162837,
+ "loss": 6.146507263183594,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.05729430541396141,
+ "learning_rate": 0.0029273189266983035,
+ "loss": 6.280686378479004,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.0410807766020298,
+ "learning_rate": 0.002926906458900473,
+ "loss": 6.246824264526367,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.044607777148485184,
+ "learning_rate": 0.0029264928535551104,
+ "loss": 6.19956111907959,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.05141043663024902,
+ "learning_rate": 0.0029260781109954515,
+ "loss": 6.118578910827637,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.029248764738440514,
+ "learning_rate": 0.0029256622315556448,
+ "loss": 6.155360221862793,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.028403814882040024,
+ "learning_rate": 0.0029252452155707582,
+ "loss": 6.097352027893066,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.025872666388750076,
+ "learning_rate": 0.002924827063376773,
+ "loss": 6.072485446929932,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.018447555601596832,
+ "learning_rate": 0.0029244077753105874,
+ "loss": 6.077045440673828,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.023166058585047722,
+ "learning_rate": 0.0029239873517100147,
+ "loss": 6.040050983428955,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.019579261541366577,
+ "learning_rate": 0.002923565792913782,
+ "loss": 6.039424896240234,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.018369073048233986,
+ "learning_rate": 0.002923143099261531,
+ "loss": 6.021180152893066,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.014188461937010288,
+ "learning_rate": 0.0029227192710938195,
+ "loss": 5.9668049812316895,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.019567398354411125,
+ "learning_rate": 0.0029222943087521182,
+ "loss": 5.989189147949219,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.014190621674060822,
+ "learning_rate": 0.0029218682125788102,
+ "loss": 5.942726135253906,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01375526562333107,
+ "learning_rate": 0.002921440982917195,
+ "loss": 5.942777633666992,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.014203300699591637,
+ "learning_rate": 0.0029210126201114836,
+ "loss": 5.927928924560547,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.014231069944798946,
+ "learning_rate": 0.0029205831245067994,
+ "loss": 5.923510551452637,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.012771515175700188,
+ "learning_rate": 0.0029201524964491796,
+ "loss": 5.908803939819336,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.013399476185441017,
+ "learning_rate": 0.0029197207362855735,
+ "loss": 5.890079021453857,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.025862423703074455,
+ "learning_rate": 0.0029192878443638427,
+ "loss": 5.919907569885254,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.011315986514091492,
+ "learning_rate": 0.0029188538210327595,
+ "loss": 5.901529788970947,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.01235947571694851,
+ "learning_rate": 0.00291841866664201,
+ "loss": 5.903595924377441,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.012175127863883972,
+ "learning_rate": 0.0029179823815421893,
+ "loss": 5.868306636810303,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01195445004850626,
+ "learning_rate": 0.0029175449660848052,
+ "loss": 5.899898529052734,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.012801481410861015,
+ "learning_rate": 0.0029171064206222745,
+ "loss": 5.890621662139893,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.011909718625247478,
+ "learning_rate": 0.002916666745507926,
+ "loss": 5.86421012878418,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01162486057728529,
+ "learning_rate": 0.0029162259410959977,
+ "loss": 5.869139671325684,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.010923407040536404,
+ "learning_rate": 0.0029157840077416374,
+ "loss": 5.869789123535156,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.011943561024963856,
+ "learning_rate": 0.002915340945800904,
+ "loss": 5.873437881469727,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.010295520536601543,
+ "learning_rate": 0.002914896755630763,
+ "loss": 5.819953918457031,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.0093373479321599,
+ "learning_rate": 0.002914451437589092,
+ "loss": 5.829481601715088,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.01182442344725132,
+ "learning_rate": 0.002914004992034674,
+ "loss": 5.855249404907227,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.00939213763922453,
+ "learning_rate": 0.0029135574193272034,
+ "loss": 5.83876371383667,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.011648556217551231,
+ "learning_rate": 0.0029131087198272813,
+ "loss": 5.824906826019287,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.011227169074118137,
+ "learning_rate": 0.002912658893896416,
+ "loss": 5.847869873046875,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.011084851808845997,
+ "learning_rate": 0.002912207941897025,
+ "loss": 5.839183807373047,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.009923825040459633,
+ "learning_rate": 0.002911755864192432,
+ "loss": 5.818821907043457,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.011300318874418736,
+ "learning_rate": 0.002911302661146867,
+ "loss": 5.819517135620117,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.009915645234286785,
+ "learning_rate": 0.0029108483331254684,
+ "loss": 5.776939392089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.008965205401182175,
+ "learning_rate": 0.00291039288049428,
+ "loss": 5.824148178100586,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.011025927029550076,
+ "learning_rate": 0.0029099363036202515,
+ "loss": 5.832704544067383,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.010143411345779896,
+ "learning_rate": 0.0029094786028712387,
+ "loss": 5.802353382110596,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.010021256282925606,
+ "learning_rate": 0.002909019778616003,
+ "loss": 5.802136421203613,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.011356706731021404,
+ "learning_rate": 0.002908559831224211,
+ "loss": 5.807962417602539,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.010525298304855824,
+ "learning_rate": 0.002908098761066433,
+ "loss": 5.787661552429199,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.011147079057991505,
+ "learning_rate": 0.0029076365685141463,
+ "loss": 5.801015377044678,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.008777810260653496,
+ "learning_rate": 0.00290717325393973,
+ "loss": 5.779858589172363,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.011423609219491482,
+ "learning_rate": 0.002906708817716469,
+ "loss": 5.777166843414307,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.008639856241643429,
+ "learning_rate": 0.00290624326021855,
+ "loss": 5.747189521789551,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.009628898464143276,
+ "learning_rate": 0.002905776581821066,
+ "loss": 5.7666916847229,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.011988377198576927,
+ "learning_rate": 0.0029053087829000097,
+ "loss": 5.777951240539551,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.01693224348127842,
+ "learning_rate": 0.00290483986383228,
+ "loss": 5.818393707275391,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.020264796912670135,
+ "learning_rate": 0.002904369824995674,
+ "loss": 5.747611045837402,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.018283747136592865,
+ "learning_rate": 0.0029038986667688964,
+ "loss": 5.789663314819336,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.825676918029785,
+ "eval_runtime": 57.6908,
+ "eval_samples_per_second": 42.329,
+ "eval_steps_per_second": 1.335,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.011447477154433727,
+ "learning_rate": 0.002903426389531549,
+ "loss": 5.724975109100342,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.012539781630039215,
+ "learning_rate": 0.002902952993664138,
+ "loss": 5.747196674346924,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.012179837562143803,
+ "learning_rate": 0.002902478479548069,
+ "loss": 5.748995780944824,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.019896456971764565,
+ "learning_rate": 0.00290200284756565,
+ "loss": 5.751996994018555,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.022261742502450943,
+ "learning_rate": 0.002901526098100089,
+ "loss": 5.7516984939575195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02683536894619465,
+ "learning_rate": 0.0029010482315354946,
+ "loss": 5.716037750244141,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.018057072535157204,
+ "learning_rate": 0.0029005692482568747,
+ "loss": 5.778608322143555,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.011646891944110394,
+ "learning_rate": 0.002900089148650138,
+ "loss": 5.7308268547058105,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02230495773255825,
+ "learning_rate": 0.002899607933102092,
+ "loss": 5.768552780151367,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.01212464738637209,
+ "learning_rate": 0.002899125602000443,
+ "loss": 5.733015060424805,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.011764907278120518,
+ "learning_rate": 0.0028986421557337973,
+ "loss": 5.748104095458984,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.01593147963285446,
+ "learning_rate": 0.002898157594691658,
+ "loss": 5.701960563659668,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.016333773732185364,
+ "learning_rate": 0.0028976719192644277,
+ "loss": 5.716808795928955,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.017547868192195892,
+ "learning_rate": 0.002897185129843406,
+ "loss": 5.724377632141113,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.01662801019847393,
+ "learning_rate": 0.002896697226820791,
+ "loss": 5.737432956695557,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.015781691297888756,
+ "learning_rate": 0.002896208210589676,
+ "loss": 5.7376203536987305,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.014363214373588562,
+ "learning_rate": 0.0028957180815440534,
+ "loss": 5.7342610359191895,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.015048054046928883,
+ "learning_rate": 0.0028952268400788115,
+ "loss": 5.715073585510254,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.013753226026892662,
+ "learning_rate": 0.002894734486589734,
+ "loss": 5.73264217376709,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.015562563203275204,
+ "learning_rate": 0.002894241021473502,
+ "loss": 5.7072038650512695,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.01520819216966629,
+ "learning_rate": 0.0028937464451276916,
+ "loss": 5.683300018310547,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.013024724088609219,
+ "learning_rate": 0.002893250757950773,
+ "loss": 5.726273536682129,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.012760856188833714,
+ "learning_rate": 0.0028927539603421128,
+ "loss": 5.7249064445495605,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.013542579486966133,
+ "learning_rate": 0.0028922560527019723,
+ "loss": 5.70849084854126,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.014801754616200924,
+ "learning_rate": 0.002891757035431506,
+ "loss": 5.725185394287109,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.013000023551285267,
+ "learning_rate": 0.0028912569089327637,
+ "loss": 5.705799102783203,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.012919292785227299,
+ "learning_rate": 0.002890755673608688,
+ "loss": 5.674324989318848,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.015904201194643974,
+ "learning_rate": 0.002890253329863115,
+ "loss": 5.680788993835449,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.015467721968889236,
+ "learning_rate": 0.0028897498781007744,
+ "loss": 5.707531929016113,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.015157260000705719,
+ "learning_rate": 0.002889245318727288,
+ "loss": 5.689894676208496,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.014204306527972221,
+ "learning_rate": 0.00288873965214917,
+ "loss": 5.713001251220703,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.022045845165848732,
+ "learning_rate": 0.0028882328787738273,
+ "loss": 5.657631874084473,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.013380175456404686,
+ "learning_rate": 0.0028877249990095573,
+ "loss": 5.702192306518555,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.016698408871889114,
+ "learning_rate": 0.0028872160132655503,
+ "loss": 5.695206642150879,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.020001472905278206,
+ "learning_rate": 0.0028867059219518866,
+ "loss": 5.731742858886719,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.027350103482604027,
+ "learning_rate": 0.0028861947254795375,
+ "loss": 5.712314605712891,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.021862946450710297,
+ "learning_rate": 0.002885682424260365,
+ "loss": 5.697043418884277,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.019216977059841156,
+ "learning_rate": 0.0028851690187071204,
+ "loss": 5.698813438415527,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.02131095714867115,
+ "learning_rate": 0.002884654509233446,
+ "loss": 5.6918864250183105,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.016583887860178947,
+ "learning_rate": 0.0028841388962538724,
+ "loss": 5.711456298828125,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.016998684033751488,
+ "learning_rate": 0.0028836221801838194,
+ "loss": 5.675747871398926,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.01705755852162838,
+ "learning_rate": 0.0028831043614395966,
+ "loss": 5.697177410125732,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.014604352414608002,
+ "learning_rate": 0.0028825854404384,
+ "loss": 5.681950092315674,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.013561096973717213,
+ "learning_rate": 0.0028820654175983162,
+ "loss": 5.67357063293457,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.021149173378944397,
+ "learning_rate": 0.0028815442933383174,
+ "loss": 5.709477424621582,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.01431041955947876,
+ "learning_rate": 0.0028810220680782644,
+ "loss": 5.7012200355529785,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.01429709978401661,
+ "learning_rate": 0.0028804987422389045,
+ "loss": 5.689794540405273,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.032025713473558426,
+ "learning_rate": 0.0028799743162418712,
+ "loss": 5.6713762283325195,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.013168714940547943,
+ "learning_rate": 0.002879448790509686,
+ "loss": 5.685478210449219,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.014088666997849941,
+ "learning_rate": 0.0028789221654657556,
+ "loss": 5.645514965057373,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.013320053927600384,
+ "learning_rate": 0.002878394441534372,
+ "loss": 5.678678512573242,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.022180087864398956,
+ "learning_rate": 0.002877865619140712,
+ "loss": 5.660525321960449,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.01271889265626669,
+ "learning_rate": 0.002877335698710839,
+ "loss": 5.652164459228516,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.011617431417107582,
+ "learning_rate": 0.002876804680671701,
+ "loss": 5.6389946937561035,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.012040726840496063,
+ "learning_rate": 0.0028762725654511282,
+ "loss": 5.628744602203369,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.013148069381713867,
+ "learning_rate": 0.0028757393534778377,
+ "loss": 5.69610595703125,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.01826276071369648,
+ "learning_rate": 0.0028752050451814276,
+ "loss": 5.634238243103027,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.020393460988998413,
+ "learning_rate": 0.0028746696409923805,
+ "loss": 5.656968593597412,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.017072148621082306,
+ "learning_rate": 0.0028741331413420627,
+ "loss": 5.668219089508057,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.01694098301231861,
+ "learning_rate": 0.0028735955466627217,
+ "loss": 5.670073509216309,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.022760670632123947,
+ "learning_rate": 0.002873056857387488,
+ "loss": 5.649288177490234,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.03613442927598953,
+ "learning_rate": 0.0028725170739503734,
+ "loss": 5.723211288452148,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.04232329502701759,
+ "learning_rate": 0.002871976196786272,
+ "loss": 5.757920742034912,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.044210851192474365,
+ "learning_rate": 0.002871434226330959,
+ "loss": 5.702426910400391,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.044886402785778046,
+ "learning_rate": 0.0028708911630210905,
+ "loss": 5.742881774902344,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.022805195301771164,
+ "learning_rate": 0.002870347007294202,
+ "loss": 5.716353416442871,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02219395525753498,
+ "learning_rate": 0.00286980175958871,
+ "loss": 5.68503475189209,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.019240057095885277,
+ "learning_rate": 0.002869255420343911,
+ "loss": 5.700656890869141,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.01957295648753643,
+ "learning_rate": 0.0028687079899999807,
+ "loss": 5.70595645904541,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016326555982232094,
+ "learning_rate": 0.0028681594689979736,
+ "loss": 5.697229385375977,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015815885737538338,
+ "learning_rate": 0.0028676098577798234,
+ "loss": 5.7020792961120605,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.015865175053477287,
+ "learning_rate": 0.002867059156788342,
+ "loss": 5.648272514343262,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.7417449951171875,
+ "eval_runtime": 51.2263,
+ "eval_samples_per_second": 47.671,
+ "eval_steps_per_second": 1.503,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.016063183546066284,
+ "learning_rate": 0.0028665073664672185,
+ "loss": 5.65411376953125,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.014323881827294827,
+ "learning_rate": 0.0028659544872610214,
+ "loss": 5.629649639129639,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.014005238190293312,
+ "learning_rate": 0.002865400519615195,
+ "loss": 5.655135631561279,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.01385372132062912,
+ "learning_rate": 0.0028648454639760615,
+ "loss": 5.632213115692139,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.014073756523430347,
+ "learning_rate": 0.0028642893207908187,
+ "loss": 5.626431465148926,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.019551943987607956,
+ "learning_rate": 0.0028637320905075406,
+ "loss": 5.595447063446045,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.017571093514561653,
+ "learning_rate": 0.002863173773575179,
+ "loss": 5.661581039428711,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.01691303588449955,
+ "learning_rate": 0.0028626143704435585,
+ "loss": 5.641129493713379,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.01584242656826973,
+ "learning_rate": 0.0028620538815633807,
+ "loss": 5.640626907348633,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.014592235907912254,
+ "learning_rate": 0.0028614923073862213,
+ "loss": 5.611204147338867,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.016636405140161514,
+ "learning_rate": 0.0028609296483645304,
+ "loss": 5.614716053009033,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.01370833721011877,
+ "learning_rate": 0.002860365904951632,
+ "loss": 5.633011817932129,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.013218311592936516,
+ "learning_rate": 0.002859801077601725,
+ "loss": 5.588958740234375,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.016971616074442863,
+ "learning_rate": 0.0028592351667698786,
+ "loss": 5.635064125061035,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.015524905174970627,
+ "learning_rate": 0.0028586681729120393,
+ "loss": 5.611454963684082,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.014769157394766808,
+ "learning_rate": 0.0028581000964850215,
+ "loss": 5.62175178527832,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.014313343912363052,
+ "learning_rate": 0.002857530937946516,
+ "loss": 5.603628158569336,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.016748322173953056,
+ "learning_rate": 0.002856960697755083,
+ "loss": 5.617765426635742,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.013503300957381725,
+ "learning_rate": 0.0028563893763701538,
+ "loss": 5.636135578155518,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.012275655753910542,
+ "learning_rate": 0.002855816974252033,
+ "loss": 5.552533149719238,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.01288773212581873,
+ "learning_rate": 0.002855243491861894,
+ "loss": 5.620527267456055,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.013687008060514927,
+ "learning_rate": 0.002854668929661781,
+ "loss": 5.618595123291016,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.015678439289331436,
+ "learning_rate": 0.0028540932881146092,
+ "loss": 5.604031562805176,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.013567988760769367,
+ "learning_rate": 0.002853516567684162,
+ "loss": 5.549149513244629,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0157297495752573,
+ "learning_rate": 0.002852938768835092,
+ "loss": 5.618505477905273,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.011867019347846508,
+ "learning_rate": 0.0028523598920329233,
+ "loss": 5.579049110412598,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.01398563478142023,
+ "learning_rate": 0.0028517799377440443,
+ "loss": 5.597678184509277,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.015525273978710175,
+ "learning_rate": 0.0028511989064357148,
+ "loss": 5.601663589477539,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.01588466763496399,
+ "learning_rate": 0.002850616798576061,
+ "loss": 5.5650434494018555,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.014276969246566296,
+ "learning_rate": 0.0028500336146340772,
+ "loss": 5.607125282287598,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.013253689743578434,
+ "learning_rate": 0.002849449355079623,
+ "loss": 5.6299824714660645,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.01176661066710949,
+ "learning_rate": 0.0028488640203834263,
+ "loss": 5.560839653015137,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.011712565086781979,
+ "learning_rate": 0.0028482776110170813,
+ "loss": 5.618312835693359,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.013515647500753403,
+ "learning_rate": 0.002847690127453047,
+ "loss": 5.566764831542969,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.020177524536848068,
+ "learning_rate": 0.0028471015701646485,
+ "loss": 5.554571628570557,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.029197178781032562,
+ "learning_rate": 0.0028465119396260757,
+ "loss": 5.6097259521484375,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.027947140857577324,
+ "learning_rate": 0.0028459212363123824,
+ "loss": 5.65598201751709,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.021089809015393257,
+ "learning_rate": 0.0028453294606994887,
+ "loss": 5.67195987701416,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.019936271011829376,
+ "learning_rate": 0.0028447366132641773,
+ "loss": 5.609482765197754,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.0176444910466671,
+ "learning_rate": 0.002844142694484094,
+ "loss": 5.60432767868042,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.023154428228735924,
+ "learning_rate": 0.0028435477048377495,
+ "loss": 5.616270065307617,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.018391480669379234,
+ "learning_rate": 0.002842951644804515,
+ "loss": 5.600588321685791,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.013716557994484901,
+ "learning_rate": 0.002842354514864627,
+ "loss": 5.595095634460449,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.014372860081493855,
+ "learning_rate": 0.002841756315499181,
+ "loss": 5.584517478942871,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.014930845238268375,
+ "learning_rate": 0.002841157047190135,
+ "loss": 5.587005138397217,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.012631502002477646,
+ "learning_rate": 0.0028405567104203095,
+ "loss": 5.594094276428223,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.014677002094686031,
+ "learning_rate": 0.0028399553056733854,
+ "loss": 5.569582939147949,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.015558961778879166,
+ "learning_rate": 0.0028393528334339025,
+ "loss": 5.596535682678223,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.014322095550596714,
+ "learning_rate": 0.0028387492941872624,
+ "loss": 5.570324420928955,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.012913414277136326,
+ "learning_rate": 0.002838144688419726,
+ "loss": 5.573002338409424,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.01251310296356678,
+ "learning_rate": 0.0028375390166184137,
+ "loss": 5.550327777862549,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.013847017660737038,
+ "learning_rate": 0.0028369322792713026,
+ "loss": 5.54567813873291,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.013789358548820019,
+ "learning_rate": 0.0028363244768672313,
+ "loss": 5.571495056152344,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.01236631814390421,
+ "learning_rate": 0.0028357156098958954,
+ "loss": 5.569730758666992,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.013210169039666653,
+ "learning_rate": 0.002835105678847847,
+ "loss": 5.53984260559082,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.01604137383401394,
+ "learning_rate": 0.0028344946842144975,
+ "loss": 5.530598163604736,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.01708815060555935,
+ "learning_rate": 0.002833882626488114,
+ "loss": 5.529256820678711,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.017222166061401367,
+ "learning_rate": 0.0028332695061618197,
+ "loss": 5.564949035644531,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.01803218014538288,
+ "learning_rate": 0.002832655323729595,
+ "loss": 5.560049057006836,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.01658225618302822,
+ "learning_rate": 0.002832040079686276,
+ "loss": 5.553183555603027,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.016834143549203873,
+ "learning_rate": 0.002831423774527554,
+ "loss": 5.573678970336914,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.01856408454477787,
+ "learning_rate": 0.002830806408749974,
+ "loss": 5.539497375488281,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.026909375563263893,
+ "learning_rate": 0.002830187982850936,
+ "loss": 5.551433563232422,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.016046348959207535,
+ "learning_rate": 0.002829568497328696,
+ "loss": 5.566728591918945,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.018582617864012718,
+ "learning_rate": 0.0028289479526823614,
+ "loss": 5.543226718902588,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.0211537666618824,
+ "learning_rate": 0.0028283263494118943,
+ "loss": 5.560711860656738,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.013894101604819298,
+ "learning_rate": 0.0028277036880181095,
+ "loss": 5.528890132904053,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.017581885680556297,
+ "learning_rate": 0.0028270799690026725,
+ "loss": 5.57473087310791,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.01622696965932846,
+ "learning_rate": 0.002826455192868104,
+ "loss": 5.578961372375488,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.01966337114572525,
+ "learning_rate": 0.0028258293601177743,
+ "loss": 5.513942718505859,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021581366658210754,
+ "learning_rate": 0.002825202471255906,
+ "loss": 5.573669910430908,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.020468758419156075,
+ "learning_rate": 0.002824574526787571,
+ "loss": 5.516103744506836,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.626922130584717,
+ "eval_runtime": 54.2404,
+ "eval_samples_per_second": 45.022,
+ "eval_steps_per_second": 1.42,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.014597049914300442,
+ "learning_rate": 0.0028239455272186936,
+ "loss": 5.506453514099121,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.020277054980397224,
+ "learning_rate": 0.002823315473056048,
+ "loss": 5.535350799560547,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.019103925675153732,
+ "learning_rate": 0.0028226843648072562,
+ "loss": 5.522078037261963,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.07158710062503815,
+ "learning_rate": 0.0028220522029807913,
+ "loss": 5.531142234802246,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01958007924258709,
+ "learning_rate": 0.0028214189880859746,
+ "loss": 5.484055042266846,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.0185184795409441,
+ "learning_rate": 0.002820784720632976,
+ "loss": 5.511264801025391,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01976991444826126,
+ "learning_rate": 0.0028201494011328133,
+ "loss": 5.51821231842041,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.02188604325056076,
+ "learning_rate": 0.002819513030097352,
+ "loss": 5.483095169067383,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.019405223429203033,
+ "learning_rate": 0.002818875608039305,
+ "loss": 5.539701461791992,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.016907615587115288,
+ "learning_rate": 0.0028182371354722306,
+ "loss": 5.518251419067383,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.02160538360476494,
+ "learning_rate": 0.0028175976129105364,
+ "loss": 5.509786605834961,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.023125721141695976,
+ "learning_rate": 0.002816957040869473,
+ "loss": 5.535561561584473,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.022233380004763603,
+ "learning_rate": 0.0028163154198651372,
+ "loss": 5.553831100463867,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.017227552831172943,
+ "learning_rate": 0.002815672750414473,
+ "loss": 5.547080993652344,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.017122473567724228,
+ "learning_rate": 0.0028150290330352665,
+ "loss": 5.493778228759766,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.020641913637518883,
+ "learning_rate": 0.0028143842682461498,
+ "loss": 5.536645889282227,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.017383966594934464,
+ "learning_rate": 0.0028137384565665967,
+ "loss": 5.506233215332031,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.017969606444239616,
+ "learning_rate": 0.002813091598516928,
+ "loss": 5.524741172790527,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.01936613954603672,
+ "learning_rate": 0.0028124436946183047,
+ "loss": 5.460989952087402,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.017492495477199554,
+ "learning_rate": 0.0028117947453927313,
+ "loss": 5.487619876861572,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.018786396831274033,
+ "learning_rate": 0.0028111447513630545,
+ "loss": 5.516829013824463,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.020149117335677147,
+ "learning_rate": 0.0028104937130529624,
+ "loss": 5.460541725158691,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.024668103083968163,
+ "learning_rate": 0.0028098416309869857,
+ "loss": 5.497687339782715,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.015888608992099762,
+ "learning_rate": 0.002809188505690494,
+ "loss": 5.494459629058838,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.019327623769640923,
+ "learning_rate": 0.0028085343376896997,
+ "loss": 5.486351490020752,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.026855193078517914,
+ "learning_rate": 0.0028078791275116532,
+ "loss": 5.509825706481934,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.026340370997786522,
+ "learning_rate": 0.002807222875684246,
+ "loss": 5.527161598205566,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.02200835384428501,
+ "learning_rate": 0.002806565582736208,
+ "loss": 5.507146835327148,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.020420534536242485,
+ "learning_rate": 0.0028059072491971085,
+ "loss": 5.5162506103515625,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02493203990161419,
+ "learning_rate": 0.0028052478755973545,
+ "loss": 5.5252556800842285,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.021293973550200462,
+ "learning_rate": 0.0028045874624681925,
+ "loss": 5.523589134216309,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.016593236476182938,
+ "learning_rate": 0.0028039260103417045,
+ "loss": 5.500586032867432,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.018857723101973534,
+ "learning_rate": 0.002803263519750811,
+ "loss": 5.52933406829834,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.015930436551570892,
+ "learning_rate": 0.0028025999912292677,
+ "loss": 5.48813533782959,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.019427936524152756,
+ "learning_rate": 0.002801935425311669,
+ "loss": 5.46573543548584,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.018123053014278412,
+ "learning_rate": 0.0028012698225334427,
+ "loss": 5.499445915222168,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.02035410702228546,
+ "learning_rate": 0.002800603183430853,
+ "loss": 5.513540267944336,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.025838350877165794,
+ "learning_rate": 0.0027999355085409996,
+ "loss": 5.565637588500977,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.029621612280607224,
+ "learning_rate": 0.002799266798401816,
+ "loss": 5.5619215965271,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.028948379680514336,
+ "learning_rate": 0.0027985970535520685,
+ "loss": 5.551331520080566,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.022135930135846138,
+ "learning_rate": 0.0027979262745313604,
+ "loss": 5.527722358703613,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.028884069994091988,
+ "learning_rate": 0.0027972544618801256,
+ "loss": 5.554599761962891,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.02499174140393734,
+ "learning_rate": 0.00279658161613963,
+ "loss": 5.543351173400879,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.021131042391061783,
+ "learning_rate": 0.0027959077378519755,
+ "loss": 5.496039390563965,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.02957489714026451,
+ "learning_rate": 0.002795232827560092,
+ "loss": 5.552300453186035,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.022832119837403297,
+ "learning_rate": 0.0027945568858077433,
+ "loss": 5.488651275634766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.02074701525270939,
+ "learning_rate": 0.002793879913139523,
+ "loss": 5.491244792938232,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.019919803366065025,
+ "learning_rate": 0.002793201910100856,
+ "loss": 5.483855247497559,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.01858169585466385,
+ "learning_rate": 0.002792522877237998,
+ "loss": 5.50787353515625,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02085857465863228,
+ "learning_rate": 0.002791842815098032,
+ "loss": 5.512872695922852,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.01407450158149004,
+ "learning_rate": 0.002791161724228872,
+ "loss": 5.472590446472168,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.01672944612801075,
+ "learning_rate": 0.002790479605179262,
+ "loss": 5.469183921813965,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.016565700992941856,
+ "learning_rate": 0.0027897964584987713,
+ "loss": 5.493677139282227,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.017017360776662827,
+ "learning_rate": 0.0027891122847377996,
+ "loss": 5.464931488037109,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.019842486828565598,
+ "learning_rate": 0.002788427084447574,
+ "loss": 5.461293697357178,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.016210904344916344,
+ "learning_rate": 0.0027877408581801467,
+ "loss": 5.454716682434082,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.02064039371907711,
+ "learning_rate": 0.002787053606488398,
+ "loss": 5.466576099395752,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.018198711797595024,
+ "learning_rate": 0.0027863653299260356,
+ "loss": 5.447747230529785,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.015937626361846924,
+ "learning_rate": 0.00278567602904759,
+ "loss": 5.5055036544799805,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.01687259040772915,
+ "learning_rate": 0.002784985704408419,
+ "loss": 5.422250747680664,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.02622046135365963,
+ "learning_rate": 0.002784294356564705,
+ "loss": 5.466894626617432,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.018672186881303787,
+ "learning_rate": 0.0027836019860734535,
+ "loss": 5.48300838470459,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.02166241593658924,
+ "learning_rate": 0.0027829085934924958,
+ "loss": 5.477658271789551,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018426023423671722,
+ "learning_rate": 0.0027822141793804856,
+ "loss": 5.487186431884766,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.022229891270399094,
+ "learning_rate": 0.0027815187442968995,
+ "loss": 5.488579273223877,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.026930972933769226,
+ "learning_rate": 0.002780822288802037,
+ "loss": 5.492000102996826,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.02733265422284603,
+ "learning_rate": 0.00278012481345702,
+ "loss": 5.4663801193237305,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.02993253618478775,
+ "learning_rate": 0.0027794263188237922,
+ "loss": 5.543368816375732,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03152153640985489,
+ "learning_rate": 0.002778726805465118,
+ "loss": 5.565644264221191,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.031395044177770615,
+ "learning_rate": 0.002778026273944581,
+ "loss": 5.553359508514404,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.026603225618600845,
+ "learning_rate": 0.002777324724826589,
+ "loss": 5.552347183227539,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.024152139201760292,
+ "learning_rate": 0.002776622158676366,
+ "loss": 5.493257522583008,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.602636814117432,
+ "eval_runtime": 54.6199,
+ "eval_samples_per_second": 44.709,
+ "eval_steps_per_second": 1.41,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.02089761756360531,
+ "learning_rate": 0.0027759185760599576,
+ "loss": 5.461747646331787,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.018181778490543365,
+ "learning_rate": 0.0027752139775442278,
+ "loss": 5.492741584777832,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.020166238769888878,
+ "learning_rate": 0.0027745083636968575,
+ "loss": 5.476003646850586,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.015491694211959839,
+ "learning_rate": 0.0027738017350863488,
+ "loss": 5.425764560699463,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.022291697561740875,
+ "learning_rate": 0.0027730940922820182,
+ "loss": 5.45866584777832,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.018074683845043182,
+ "learning_rate": 0.002772385435854002,
+ "loss": 5.465143203735352,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.016872268170118332,
+ "learning_rate": 0.002771675766373251,
+ "loss": 5.446003437042236,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.023103339597582817,
+ "learning_rate": 0.0027709650844115336,
+ "loss": 5.411036491394043,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.027957158163189888,
+ "learning_rate": 0.0027702533905414324,
+ "loss": 5.452442169189453,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.019935399293899536,
+ "learning_rate": 0.002769540685336348,
+ "loss": 5.412145614624023,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.014351371675729752,
+ "learning_rate": 0.0027688269693704926,
+ "loss": 5.434524059295654,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.0152254244312644,
+ "learning_rate": 0.0027681122432188948,
+ "loss": 5.428544998168945,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.015071330592036247,
+ "learning_rate": 0.0027673965074573965,
+ "loss": 5.459929466247559,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.023415958508849144,
+ "learning_rate": 0.0027666797626626537,
+ "loss": 5.447159767150879,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.017228802666068077,
+ "learning_rate": 0.0027659620094121337,
+ "loss": 5.442273139953613,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.014184504747390747,
+ "learning_rate": 0.002765243248284118,
+ "loss": 5.448305130004883,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.024384526535868645,
+ "learning_rate": 0.0027645234798576988,
+ "loss": 5.4076337814331055,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.043152231723070145,
+ "learning_rate": 0.0027638027047127813,
+ "loss": 5.45344352722168,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.02889758162200451,
+ "learning_rate": 0.002763080923430081,
+ "loss": 5.489194393157959,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.02686968818306923,
+ "learning_rate": 0.002762358136591123,
+ "loss": 5.493659496307373,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.020467868074774742,
+ "learning_rate": 0.002761634344778245,
+ "loss": 5.410489559173584,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.019759224727749825,
+ "learning_rate": 0.002760909548574592,
+ "loss": 5.439385890960693,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.0233843382447958,
+ "learning_rate": 0.0027601837485641194,
+ "loss": 5.445036888122559,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.031753458082675934,
+ "learning_rate": 0.0027594569453315916,
+ "loss": 5.503895282745361,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.02089349366724491,
+ "learning_rate": 0.002758729139462581,
+ "loss": 5.45689582824707,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.03339357674121857,
+ "learning_rate": 0.002758000331543466,
+ "loss": 5.522361755371094,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019328996539115906,
+ "learning_rate": 0.002757270522161436,
+ "loss": 5.441373825073242,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019863856956362724,
+ "learning_rate": 0.002756539711904484,
+ "loss": 5.4633283615112305,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.02141754701733589,
+ "learning_rate": 0.0027558079013614116,
+ "loss": 5.453215599060059,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016591915860772133,
+ "learning_rate": 0.002755075091121824,
+ "loss": 5.463696479797363,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.01955028623342514,
+ "learning_rate": 0.0027543412817761345,
+ "loss": 5.411355018615723,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.016791662201285362,
+ "learning_rate": 0.0027536064739155596,
+ "loss": 5.459614276885986,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.019661981612443924,
+ "learning_rate": 0.00275287066813212,
+ "loss": 5.440699577331543,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.01927965320646763,
+ "learning_rate": 0.002752133865018642,
+ "loss": 5.409865856170654,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.016483580693602562,
+ "learning_rate": 0.002751396065168754,
+ "loss": 5.390999794006348,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016264773905277252,
+ "learning_rate": 0.0027506572691768888,
+ "loss": 5.418543815612793,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.01958460919559002,
+ "learning_rate": 0.00274991747763828,
+ "loss": 5.403289794921875,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.01861695572733879,
+ "learning_rate": 0.002749176691148965,
+ "loss": 5.431911468505859,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.029594263061881065,
+ "learning_rate": 0.002748434910305782,
+ "loss": 5.399242401123047,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.028574613854289055,
+ "learning_rate": 0.0027476921357063697,
+ "loss": 5.440446853637695,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.028612585738301277,
+ "learning_rate": 0.0027469483679491693,
+ "loss": 5.410346031188965,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.020704898983240128,
+ "learning_rate": 0.00274620360763342,
+ "loss": 5.44694709777832,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.020542364567518234,
+ "learning_rate": 0.002745457855359163,
+ "loss": 5.402653694152832,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.015762662515044212,
+ "learning_rate": 0.0027447111117272363,
+ "loss": 5.396594047546387,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022096984088420868,
+ "learning_rate": 0.002743963377339278,
+ "loss": 5.398744106292725,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.018791764974594116,
+ "learning_rate": 0.002743214652797724,
+ "loss": 5.397758483886719,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01642526686191559,
+ "learning_rate": 0.0027424649387058093,
+ "loss": 5.411459922790527,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.017931703478097916,
+ "learning_rate": 0.0027417142356675637,
+ "loss": 5.390264511108398,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018819602206349373,
+ "learning_rate": 0.0027409625442878153,
+ "loss": 5.407539367675781,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.019389411434531212,
+ "learning_rate": 0.0027402098651721886,
+ "loss": 5.386474609375,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.013645983301103115,
+ "learning_rate": 0.002739456198927104,
+ "loss": 5.373845100402832,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.013846895657479763,
+ "learning_rate": 0.002738701546159776,
+ "loss": 5.387155055999756,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013553125783801079,
+ "learning_rate": 0.002737945907478215,
+ "loss": 5.377801895141602,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.012771612964570522,
+ "learning_rate": 0.002737189283491225,
+ "loss": 5.344066143035889,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.020465301349759102,
+ "learning_rate": 0.002736431674808405,
+ "loss": 5.382771968841553,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.014798106625676155,
+ "learning_rate": 0.0027356730820401465,
+ "loss": 5.370522499084473,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.01566634140908718,
+ "learning_rate": 0.0027349135057976332,
+ "loss": 5.351980209350586,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.017485102638602257,
+ "learning_rate": 0.002734152946692843,
+ "loss": 5.355443477630615,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.016646096482872963,
+ "learning_rate": 0.0027333914053385444,
+ "loss": 5.34829044342041,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.017293352633714676,
+ "learning_rate": 0.0027326288823482964,
+ "loss": 5.3519487380981445,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.018873002380132675,
+ "learning_rate": 0.002731865378336452,
+ "loss": 5.348442077636719,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.015504784882068634,
+ "learning_rate": 0.002731100893918151,
+ "loss": 5.341490268707275,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.014270870015025139,
+ "learning_rate": 0.0027303354297093254,
+ "loss": 5.359251022338867,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.018820680677890778,
+ "learning_rate": 0.002729568986326695,
+ "loss": 5.390280723571777,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.021008187904953957,
+ "learning_rate": 0.0027288015643877708,
+ "loss": 5.338380336761475,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.016266655176877975,
+ "learning_rate": 0.0027280331645108498,
+ "loss": 5.347043037414551,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.01831045374274254,
+ "learning_rate": 0.0027272637873150185,
+ "loss": 5.326537132263184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02073794975876808,
+ "learning_rate": 0.002726493433420149,
+ "loss": 5.3607025146484375,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.022534754127264023,
+ "learning_rate": 0.0027257221034469033,
+ "loss": 5.353292465209961,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.021627362817525864,
+ "learning_rate": 0.002724949798016727,
+ "loss": 5.373323917388916,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.018226243555545807,
+ "learning_rate": 0.0027241765177518523,
+ "loss": 5.344578266143799,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.02032875455915928,
+ "learning_rate": 0.002723402263275298,
+ "loss": 5.329159259796143,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.44969367980957,
+ "eval_runtime": 51.2294,
+ "eval_samples_per_second": 47.668,
+ "eval_steps_per_second": 1.503,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.01613701693713665,
+ "learning_rate": 0.0027226270352108664,
+ "loss": 5.3024678230285645,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01940973661839962,
+ "learning_rate": 0.0027218508341831445,
+ "loss": 5.270181179046631,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.015840493142604828,
+ "learning_rate": 0.002721073660817504,
+ "loss": 5.261706352233887,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.017390675842761993,
+ "learning_rate": 0.0027202955157401,
+ "loss": 5.262207984924316,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.015184012241661549,
+ "learning_rate": 0.0027195163995778686,
+ "loss": 5.297155380249023,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.013417073525488377,
+ "learning_rate": 0.00271873631295853,
+ "loss": 5.279838562011719,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.013024317100644112,
+ "learning_rate": 0.0027179552565105864,
+ "loss": 5.265446186065674,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.01426338218152523,
+ "learning_rate": 0.002717173230863321,
+ "loss": 5.284180164337158,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.019604040309786797,
+ "learning_rate": 0.002716390236646797,
+ "loss": 5.286281108856201,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.019039476290345192,
+ "learning_rate": 0.0027156062744918587,
+ "loss": 5.290433406829834,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.019284334033727646,
+ "learning_rate": 0.0027148213450301313,
+ "loss": 5.262433052062988,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.0145154669880867,
+ "learning_rate": 0.0027140354488940175,
+ "loss": 5.24387264251709,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.014776417054235935,
+ "learning_rate": 0.0027132485867166992,
+ "loss": 5.282435417175293,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.012711729854345322,
+ "learning_rate": 0.0027124607591321377,
+ "loss": 5.246735572814941,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.013394835405051708,
+ "learning_rate": 0.0027116719667750716,
+ "loss": 5.270739555358887,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.013468850404024124,
+ "learning_rate": 0.002710882210281016,
+ "loss": 5.27161169052124,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.030262840911746025,
+ "learning_rate": 0.002710091490286264,
+ "loss": 5.256673812866211,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.013560288585722446,
+ "learning_rate": 0.0027092998074278846,
+ "loss": 5.253674507141113,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014592813327908516,
+ "learning_rate": 0.002708507162343721,
+ "loss": 5.25804328918457,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014328169636428356,
+ "learning_rate": 0.0027077135556723953,
+ "loss": 5.235117435455322,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.01277366653084755,
+ "learning_rate": 0.0027069189880533,
+ "loss": 5.2053985595703125,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.012948785908520222,
+ "learning_rate": 0.0027061234601266043,
+ "loss": 5.214653015136719,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.014892828650772572,
+ "learning_rate": 0.002705326972533252,
+ "loss": 5.227547645568848,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01427230704575777,
+ "learning_rate": 0.002704529525914957,
+ "loss": 5.241065502166748,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.01650116965174675,
+ "learning_rate": 0.0027037311209142092,
+ "loss": 5.227713108062744,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.019728578627109528,
+ "learning_rate": 0.0027029317581742677,
+ "loss": 5.2624921798706055,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.023847149685025215,
+ "learning_rate": 0.0027021314383391657,
+ "loss": 5.278811454772949,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.020690469071269035,
+ "learning_rate": 0.0027013301620537056,
+ "loss": 5.238919258117676,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.021211959421634674,
+ "learning_rate": 0.0027005279299634616,
+ "loss": 5.253684997558594,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.025854922831058502,
+ "learning_rate": 0.002699724742714778,
+ "loss": 5.2538652420043945,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.0215485580265522,
+ "learning_rate": 0.0026989206009547678,
+ "loss": 5.224540710449219,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.019827792420983315,
+ "learning_rate": 0.0026981155053313137,
+ "loss": 5.220674514770508,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.017689093947410583,
+ "learning_rate": 0.002697309456493066,
+ "loss": 5.228511810302734,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.022564806044101715,
+ "learning_rate": 0.002696502455089445,
+ "loss": 5.295933723449707,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.020070280879735947,
+ "learning_rate": 0.0026956945017706357,
+ "loss": 5.230920791625977,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.025744661688804626,
+ "learning_rate": 0.002694885597187593,
+ "loss": 5.296453952789307,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.021263007074594498,
+ "learning_rate": 0.0026940757419920357,
+ "loss": 5.220481872558594,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.017577245831489563,
+ "learning_rate": 0.0026932649368364495,
+ "loss": 5.231488227844238,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.015562832355499268,
+ "learning_rate": 0.002692453182374086,
+ "loss": 5.175224304199219,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.014115403406322002,
+ "learning_rate": 0.0026916404792589606,
+ "loss": 5.228387832641602,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.012863830663263798,
+ "learning_rate": 0.002690826828145854,
+ "loss": 5.186727523803711,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.014151126146316528,
+ "learning_rate": 0.002690012229690309,
+ "loss": 5.1694817543029785,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.01316932775080204,
+ "learning_rate": 0.0026891966845486353,
+ "loss": 5.19931697845459,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.011340726166963577,
+ "learning_rate": 0.0026883801933779003,
+ "loss": 5.238619804382324,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.012616089545190334,
+ "learning_rate": 0.0026875627568359378,
+ "loss": 5.183376789093018,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.012274038046598434,
+ "learning_rate": 0.002686744375581341,
+ "loss": 5.199946403503418,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.014990398660302162,
+ "learning_rate": 0.002685925050273465,
+ "loss": 5.168451309204102,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.011399757117033005,
+ "learning_rate": 0.0026851047815724255,
+ "loss": 5.173828125,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.013638390228152275,
+ "learning_rate": 0.0026842835701390982,
+ "loss": 5.142702102661133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.014472718350589275,
+ "learning_rate": 0.0026834614166351177,
+ "loss": 5.173436164855957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.014509808272123337,
+ "learning_rate": 0.002682638321722879,
+ "loss": 5.173975467681885,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.013576583005487919,
+ "learning_rate": 0.0026818142860655352,
+ "loss": 5.1435089111328125,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.012914472259581089,
+ "learning_rate": 0.0026809893103269954,
+ "loss": 5.151381492614746,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.013803352601826191,
+ "learning_rate": 0.0026801633951719288,
+ "loss": 5.133159637451172,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.018488213419914246,
+ "learning_rate": 0.0026793365412657606,
+ "loss": 5.183903694152832,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.01843969337642193,
+ "learning_rate": 0.002678508749274671,
+ "loss": 5.1660356521606445,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.017181839793920517,
+ "learning_rate": 0.0026776800198655983,
+ "loss": 5.179847717285156,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0148160420358181,
+ "learning_rate": 0.0026768503537062333,
+ "loss": 5.149198532104492,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.013251296244561672,
+ "learning_rate": 0.0026760197514650247,
+ "loss": 5.192000389099121,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.010746927931904793,
+ "learning_rate": 0.0026751882138111717,
+ "loss": 5.138680458068848,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.012316994369029999,
+ "learning_rate": 0.002674355741414631,
+ "loss": 5.169965744018555,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.01191797386854887,
+ "learning_rate": 0.0026735223349461106,
+ "loss": 5.128267288208008,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.013525084592401981,
+ "learning_rate": 0.0026726879950770698,
+ "loss": 5.124795436859131,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.014843370765447617,
+ "learning_rate": 0.002671852722479721,
+ "loss": 5.1497087478637695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.015285412780940533,
+ "learning_rate": 0.00267101651782703,
+ "loss": 5.093118190765381,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.013555832207202911,
+ "learning_rate": 0.0026701793817927107,
+ "loss": 5.13249397277832,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.010569742880761623,
+ "learning_rate": 0.0026693413150512273,
+ "loss": 5.084077835083008,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.011483813636004925,
+ "learning_rate": 0.0026685023182777975,
+ "loss": 5.128049850463867,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.013724366202950478,
+ "learning_rate": 0.0026676623921483844,
+ "loss": 5.157539367675781,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.014337432570755482,
+ "learning_rate": 0.0026668215373397016,
+ "loss": 5.149948596954346,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.014101454056799412,
+ "learning_rate": 0.00266597975452921,
+ "loss": 5.117250442504883,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.016004368662834167,
+ "learning_rate": 0.0026651370443951197,
+ "loss": 5.143221855163574,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.249833583831787,
+ "eval_runtime": 51.5543,
+ "eval_samples_per_second": 47.368,
+ "eval_steps_per_second": 1.494,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.01297403872013092,
+ "learning_rate": 0.0026642934076163867,
+ "loss": 5.089287281036377,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.012867718935012817,
+ "learning_rate": 0.0026634488448727144,
+ "loss": 5.0784101486206055,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.014600927010178566,
+ "learning_rate": 0.0026626033568445504,
+ "loss": 5.050220489501953,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.016619572415947914,
+ "learning_rate": 0.0026617569442130905,
+ "loss": 5.079303741455078,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.015560587868094444,
+ "learning_rate": 0.0026609096076602736,
+ "loss": 5.072385311126709,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.01774057187139988,
+ "learning_rate": 0.0026600613478687833,
+ "loss": 5.122529983520508,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.019971400499343872,
+ "learning_rate": 0.0026592121655220472,
+ "loss": 5.085966110229492,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.017667256295681,
+ "learning_rate": 0.0026583620613042365,
+ "loss": 5.078432559967041,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.022369660437107086,
+ "learning_rate": 0.002657511035900264,
+ "loss": 5.0630903244018555,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.017964458093047142,
+ "learning_rate": 0.002656659089995787,
+ "loss": 5.0846076011657715,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.021769145503640175,
+ "learning_rate": 0.002655806224277201,
+ "loss": 5.081344127655029,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.02487620897591114,
+ "learning_rate": 0.0026549524394316464,
+ "loss": 5.106287956237793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.022516218945384026,
+ "learning_rate": 0.0026540977361470012,
+ "loss": 5.128098487854004,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.02099926583468914,
+ "learning_rate": 0.0026532421151118845,
+ "loss": 5.103562355041504,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.01899387687444687,
+ "learning_rate": 0.002652385577015655,
+ "loss": 5.062675476074219,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.016801996156573296,
+ "learning_rate": 0.002651528122548409,
+ "loss": 5.096002578735352,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.014009200967848301,
+ "learning_rate": 0.0026506697524009836,
+ "loss": 5.075201034545898,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.014653858728706837,
+ "learning_rate": 0.002649810467264951,
+ "loss": 5.090478897094727,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.014918150380253792,
+ "learning_rate": 0.0026489502678326226,
+ "loss": 5.070797920227051,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.01327288057655096,
+ "learning_rate": 0.0026480891547970443,
+ "loss": 5.030500411987305,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.012885140255093575,
+ "learning_rate": 0.002647227128852,
+ "loss": 5.0845537185668945,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.012848512269556522,
+ "learning_rate": 0.002646364190692009,
+ "loss": 5.079484939575195,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.01149789709597826,
+ "learning_rate": 0.002645500341012324,
+ "loss": 5.036316394805908,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.010360757820308208,
+ "learning_rate": 0.0026446355805089337,
+ "loss": 5.04129695892334,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.010426290333271027,
+ "learning_rate": 0.00264376990987856,
+ "loss": 5.051270484924316,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.01018153689801693,
+ "learning_rate": 0.002642903329818658,
+ "loss": 5.102342128753662,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.011879283003509045,
+ "learning_rate": 0.0026420358410274153,
+ "loss": 5.053501129150391,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.010418329387903214,
+ "learning_rate": 0.002641167444203752,
+ "loss": 5.030653476715088,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.012302013114094734,
+ "learning_rate": 0.0026402981400473202,
+ "loss": 5.0402069091796875,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.011392916552722454,
+ "learning_rate": 0.002639427929258502,
+ "loss": 5.024111747741699,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.013075058348476887,
+ "learning_rate": 0.0026385568125384114,
+ "loss": 5.021731376647949,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015349832363426685,
+ "learning_rate": 0.0026376847905888903,
+ "loss": 5.03192138671875,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01588309183716774,
+ "learning_rate": 0.0026368118641125113,
+ "loss": 4.966363906860352,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.014378590509295464,
+ "learning_rate": 0.0026359380338125764,
+ "loss": 5.070773124694824,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.012306655757129192,
+ "learning_rate": 0.002635063300393114,
+ "loss": 5.025704860687256,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01212004292756319,
+ "learning_rate": 0.002634187664558881,
+ "loss": 4.965104103088379,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013538849540054798,
+ "learning_rate": 0.0026333111270153624,
+ "loss": 5.029810905456543,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.012052906677126884,
+ "learning_rate": 0.0026324336884687672,
+ "loss": 5.0262908935546875,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.011368660256266594,
+ "learning_rate": 0.002631555349626033,
+ "loss": 5.001485824584961,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.010215395130217075,
+ "learning_rate": 0.0026306761111948205,
+ "loss": 5.019739151000977,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.011449878104031086,
+ "learning_rate": 0.0026297959738835167,
+ "loss": 4.9816083908081055,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.012381811626255512,
+ "learning_rate": 0.0026289149384012326,
+ "loss": 5.025032043457031,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.00974641740322113,
+ "learning_rate": 0.002628033005457802,
+ "loss": 5.000177383422852,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.009515874087810516,
+ "learning_rate": 0.0026271501757637833,
+ "loss": 4.996979713439941,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.011125455610454082,
+ "learning_rate": 0.0026262664500304555,
+ "loss": 4.996796607971191,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.013290468603372574,
+ "learning_rate": 0.0026253818289698207,
+ "loss": 5.053703308105469,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.014650283381342888,
+ "learning_rate": 0.002624496313294602,
+ "loss": 5.001321792602539,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014537827111780643,
+ "learning_rate": 0.002623609903718244,
+ "loss": 5.024584770202637,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.012908566743135452,
+ "learning_rate": 0.0026227226009549105,
+ "loss": 5.042878150939941,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.01446808222681284,
+ "learning_rate": 0.002621834405719485,
+ "loss": 5.020857810974121,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.01535153016448021,
+ "learning_rate": 0.0026209453187275708,
+ "loss": 5.0056304931640625,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01815616339445114,
+ "learning_rate": 0.0026200553406954892,
+ "loss": 5.0467376708984375,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.023135516792535782,
+ "learning_rate": 0.0026191644723402795,
+ "loss": 5.033995628356934,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.019682930782437325,
+ "learning_rate": 0.002618272714379698,
+ "loss": 5.008149147033691,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.018521904945373535,
+ "learning_rate": 0.002617380067532217,
+ "loss": 5.007076263427734,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.02118663489818573,
+ "learning_rate": 0.0026164865325170278,
+ "loss": 5.05519962310791,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024684788659214973,
+ "learning_rate": 0.0026155921100540344,
+ "loss": 5.027571201324463,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.02609819918870926,
+ "learning_rate": 0.0026146968008638567,
+ "loss": 5.10361385345459,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.025958053767681122,
+ "learning_rate": 0.00261380060566783,
+ "loss": 5.0911455154418945,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.017067518085241318,
+ "learning_rate": 0.002612903525188002,
+ "loss": 5.045783042907715,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.016510486602783203,
+ "learning_rate": 0.0026120055601471345,
+ "loss": 5.069464683532715,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015552333556115627,
+ "learning_rate": 0.0026111067112687012,
+ "loss": 5.024248123168945,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.013432685285806656,
+ "learning_rate": 0.0026102069792768886,
+ "loss": 5.004837989807129,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.012300201691687107,
+ "learning_rate": 0.0026093063648965943,
+ "loss": 5.031093597412109,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.01197744719684124,
+ "learning_rate": 0.002608404868853428,
+ "loss": 4.965237617492676,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.01100246049463749,
+ "learning_rate": 0.002607502491873708,
+ "loss": 5.04719352722168,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.010111447423696518,
+ "learning_rate": 0.0026065992346844626,
+ "loss": 4.986410617828369,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.010669790208339691,
+ "learning_rate": 0.002605695098013431,
+ "loss": 4.9825663566589355,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.010805477388203144,
+ "learning_rate": 0.0026047900825890593,
+ "loss": 4.991506576538086,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.010886074975132942,
+ "learning_rate": 0.002603884189140502,
+ "loss": 4.987531661987305,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.010735839605331421,
+ "learning_rate": 0.0026029774183976206,
+ "loss": 4.982756614685059,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.011092593893408775,
+ "learning_rate": 0.002602069771090985,
+ "loss": 4.994824409484863,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 5.13364315032959,
+ "eval_runtime": 50.4195,
+ "eval_samples_per_second": 48.434,
+ "eval_steps_per_second": 1.527,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.01005905494093895,
+ "learning_rate": 0.0026011612479518694,
+ "loss": 4.918412685394287,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.009904682636260986,
+ "learning_rate": 0.002600251849712255,
+ "loss": 4.924630165100098,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.009543103165924549,
+ "learning_rate": 0.0025993415771048273,
+ "loss": 4.940309524536133,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.011031470261514187,
+ "learning_rate": 0.0025984304308629765,
+ "loss": 4.924609184265137,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.009511022828519344,
+ "learning_rate": 0.002597518411720796,
+ "loss": 4.9102325439453125,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.00886621605604887,
+ "learning_rate": 0.0025966055204130847,
+ "loss": 4.9236907958984375,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.010057039558887482,
+ "learning_rate": 0.002595691757675341,
+ "loss": 4.908633232116699,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.01098555512726307,
+ "learning_rate": 0.0025947771242437684,
+ "loss": 4.935927391052246,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.008800056762993336,
+ "learning_rate": 0.0025938616208552694,
+ "loss": 4.873291969299316,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.009747766889631748,
+ "learning_rate": 0.0025929452482474497,
+ "loss": 4.954436302185059,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.010785684920847416,
+ "learning_rate": 0.002592028007158613,
+ "loss": 4.924927711486816,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.012620717287063599,
+ "learning_rate": 0.0025911098983277645,
+ "loss": 4.953606605529785,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.014428738504648209,
+ "learning_rate": 0.002590190922494607,
+ "loss": 4.951706886291504,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.014496048912405968,
+ "learning_rate": 0.0025892710803995444,
+ "loss": 4.941249847412109,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.0109238987788558,
+ "learning_rate": 0.0025883503727836764,
+ "loss": 4.939975261688232,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.012271072715520859,
+ "learning_rate": 0.0025874288003887996,
+ "loss": 4.880942344665527,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.016302887350320816,
+ "learning_rate": 0.0025865063639574086,
+ "loss": 4.931602478027344,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.016137251630425453,
+ "learning_rate": 0.0025855830642326945,
+ "loss": 4.918013095855713,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.014732961542904377,
+ "learning_rate": 0.002584658901958542,
+ "loss": 4.927577495574951,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.016387946903705597,
+ "learning_rate": 0.0025837338778795325,
+ "loss": 4.930819034576416,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.017146151512861252,
+ "learning_rate": 0.002582807992740941,
+ "loss": 4.928377151489258,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.016153613105416298,
+ "learning_rate": 0.0025818812472887375,
+ "loss": 4.975607395172119,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.013386047445237637,
+ "learning_rate": 0.0025809536422695825,
+ "loss": 4.9434709548950195,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.012328253127634525,
+ "learning_rate": 0.0025800251784308314,
+ "loss": 4.917585372924805,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.012000586837530136,
+ "learning_rate": 0.0025790958565205316,
+ "loss": 4.9137725830078125,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.009463997557759285,
+ "learning_rate": 0.002578165677287419,
+ "loss": 4.910362243652344,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.010591360740363598,
+ "learning_rate": 0.0025772346414809242,
+ "loss": 4.916032314300537,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.010429762303829193,
+ "learning_rate": 0.002576302749851164,
+ "loss": 4.934561729431152,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.12526583671569824,
+ "learning_rate": 0.0025753700031489486,
+ "loss": 4.897515773773193,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01368892751634121,
+ "learning_rate": 0.0025744364021257743,
+ "loss": 4.94814920425415,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.016603967174887657,
+ "learning_rate": 0.0025735019475338254,
+ "loss": 4.8996758460998535,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.06110106781125069,
+ "learning_rate": 0.0025725666401259775,
+ "loss": 4.946794509887695,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.018292423337697983,
+ "learning_rate": 0.002571630480655789,
+ "loss": 4.958992958068848,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.01852245442569256,
+ "learning_rate": 0.0025706934698775073,
+ "loss": 4.942556381225586,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.0195649154484272,
+ "learning_rate": 0.0025697556085460646,
+ "loss": 4.895773887634277,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.017820244655013084,
+ "learning_rate": 0.00256881689741708,
+ "loss": 4.915250778198242,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.016814742237329483,
+ "learning_rate": 0.002567877337246855,
+ "loss": 4.9521260261535645,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.015319781377911568,
+ "learning_rate": 0.002566936928792377,
+ "loss": 4.899402141571045,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.01607401855289936,
+ "learning_rate": 0.0025659956728113146,
+ "loss": 4.909913063049316,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.015564955770969391,
+ "learning_rate": 0.002565053570062023,
+ "loss": 4.947719573974609,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.017502037808299065,
+ "learning_rate": 0.002564110621303536,
+ "loss": 4.95107364654541,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.014791073277592659,
+ "learning_rate": 0.002563166827295571,
+ "loss": 4.8995866775512695,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.012700124643743038,
+ "learning_rate": 0.002562222188798525,
+ "loss": 4.922100067138672,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.011235272511839867,
+ "learning_rate": 0.0025612767065734763,
+ "loss": 4.884456634521484,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.013009768910706043,
+ "learning_rate": 0.002560330381382183,
+ "loss": 4.869503021240234,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.0102638378739357,
+ "learning_rate": 0.0025593832139870834,
+ "loss": 4.934551239013672,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.011081870645284653,
+ "learning_rate": 0.0025584352051512913,
+ "loss": 4.937268257141113,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.010681218467652798,
+ "learning_rate": 0.002557486355638601,
+ "loss": 4.920559883117676,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.010962890461087227,
+ "learning_rate": 0.002556536666213484,
+ "loss": 4.936082363128662,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.010790379717946053,
+ "learning_rate": 0.0025555861376410868,
+ "loss": 4.9506449699401855,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.010659157298505306,
+ "learning_rate": 0.0025546347706872346,
+ "loss": 4.872883319854736,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.012962655164301395,
+ "learning_rate": 0.0025536825661184254,
+ "loss": 4.901210784912109,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.014810293912887573,
+ "learning_rate": 0.002552729524701833,
+ "loss": 4.895804405212402,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.010993832722306252,
+ "learning_rate": 0.0025517756472053068,
+ "loss": 4.891035079956055,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.013413915410637856,
+ "learning_rate": 0.002550820934397368,
+ "loss": 4.900317192077637,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.01580147072672844,
+ "learning_rate": 0.0025498653870472117,
+ "loss": 4.883669853210449,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01864512264728546,
+ "learning_rate": 0.0025489090059247045,
+ "loss": 4.881938457489014,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017185743898153305,
+ "learning_rate": 0.0025479517918003862,
+ "loss": 4.853325843811035,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.012045156210660934,
+ "learning_rate": 0.002546993745445466,
+ "loss": 4.9152326583862305,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.013420728035271168,
+ "learning_rate": 0.0025460348676318257,
+ "loss": 4.895130157470703,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.011561610735952854,
+ "learning_rate": 0.0025450751591320152,
+ "loss": 4.890257835388184,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013164008036255836,
+ "learning_rate": 0.0025441146207192547,
+ "loss": 4.885972023010254,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01120204571634531,
+ "learning_rate": 0.002543153253167432,
+ "loss": 4.882246971130371,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.011342627927660942,
+ "learning_rate": 0.002542191057251104,
+ "loss": 4.862927436828613,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.010852213948965073,
+ "learning_rate": 0.0025412280337454942,
+ "loss": 4.879744529724121,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.012851282022893429,
+ "learning_rate": 0.0025402641834264936,
+ "loss": 4.882238388061523,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.011839655227959156,
+ "learning_rate": 0.002539299507070659,
+ "loss": 4.85658073425293,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.011315262876451015,
+ "learning_rate": 0.002538334005455213,
+ "loss": 4.9234771728515625,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.011661105789244175,
+ "learning_rate": 0.0025373676793580416,
+ "loss": 4.855696678161621,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.009712521918118,
+ "learning_rate": 0.0025364005295576975,
+ "loss": 4.89942741394043,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.011002928018569946,
+ "learning_rate": 0.0025354325568333956,
+ "loss": 4.871459007263184,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.01509513333439827,
+ "learning_rate": 0.0025344637619650135,
+ "loss": 4.834778785705566,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 5.031810283660889,
+ "eval_runtime": 51.9412,
+ "eval_samples_per_second": 47.015,
+ "eval_steps_per_second": 1.482,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.010317740961909294,
+ "learning_rate": 0.0025334941457330925,
+ "loss": 4.8169708251953125,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.00892715621739626,
+ "learning_rate": 0.0025325237089188343,
+ "loss": 4.852178573608398,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.011608307249844074,
+ "learning_rate": 0.002531552452304103,
+ "loss": 4.790487289428711,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.013100114651024342,
+ "learning_rate": 0.0025305803766714216,
+ "loss": 4.863536834716797,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.016571057960391045,
+ "learning_rate": 0.002529607482803974,
+ "loss": 4.840987205505371,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.014595622196793556,
+ "learning_rate": 0.002528633771485604,
+ "loss": 4.815027236938477,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.014603931456804276,
+ "learning_rate": 0.002527659243500812,
+ "loss": 4.853024959564209,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.019107041880488396,
+ "learning_rate": 0.0025266838996347595,
+ "loss": 4.829808235168457,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.020292190834879875,
+ "learning_rate": 0.0025257077406732613,
+ "loss": 4.775232315063477,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.014630074612796307,
+ "learning_rate": 0.0025247307674027926,
+ "loss": 4.826682090759277,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.019507668912410736,
+ "learning_rate": 0.0025237529806104823,
+ "loss": 4.843222141265869,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.023950330913066864,
+ "learning_rate": 0.002522774381084116,
+ "loss": 4.820512771606445,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.022547820582985878,
+ "learning_rate": 0.002521794969612133,
+ "loss": 4.854910850524902,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.017819277942180634,
+ "learning_rate": 0.0025208147469836283,
+ "loss": 4.825885772705078,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.017989592626690865,
+ "learning_rate": 0.002519833713988349,
+ "loss": 4.8357014656066895,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.013516640290617943,
+ "learning_rate": 0.0025188518714166956,
+ "loss": 4.83729362487793,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.015956101939082146,
+ "learning_rate": 0.002517869220059721,
+ "loss": 4.807845592498779,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.013967473991215229,
+ "learning_rate": 0.002516885760709131,
+ "loss": 4.828944206237793,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.01270931027829647,
+ "learning_rate": 0.0025159014941572785,
+ "loss": 4.826594352722168,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.012433716095983982,
+ "learning_rate": 0.002514916421197171,
+ "loss": 4.797939777374268,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.011882808059453964,
+ "learning_rate": 0.0025139305426224636,
+ "loss": 4.78917932510376,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.011193005368113518,
+ "learning_rate": 0.0025129438592274614,
+ "loss": 4.787964820861816,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.011873407289385796,
+ "learning_rate": 0.002511956371807116,
+ "loss": 4.81754732131958,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.011407418176531792,
+ "learning_rate": 0.002510968081157029,
+ "loss": 4.781639575958252,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.009560304693877697,
+ "learning_rate": 0.002509978988073449,
+ "loss": 4.8163957595825195,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.009969786740839481,
+ "learning_rate": 0.0025089890933532697,
+ "loss": 4.795045852661133,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.010739767923951149,
+ "learning_rate": 0.0025079983977940312,
+ "loss": 4.847584247589111,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.009999158792197704,
+ "learning_rate": 0.0025070069021939194,
+ "loss": 4.834909439086914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.009302482940256596,
+ "learning_rate": 0.002506014607351765,
+ "loss": 4.808745384216309,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.010687744244933128,
+ "learning_rate": 0.0025050215140670403,
+ "loss": 4.789101600646973,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.010930616408586502,
+ "learning_rate": 0.0025040276231398637,
+ "loss": 4.817464351654053,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.013101065531373024,
+ "learning_rate": 0.0025030329353709956,
+ "loss": 4.827780723571777,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.01223307941108942,
+ "learning_rate": 0.0025020374515618373,
+ "loss": 4.8367414474487305,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.013091634958982468,
+ "learning_rate": 0.0025010411725144323,
+ "loss": 4.839454650878906,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.012114843353629112,
+ "learning_rate": 0.0025000440990314643,
+ "loss": 4.832447052001953,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014664588496088982,
+ "learning_rate": 0.0024990462319162588,
+ "loss": 4.740812301635742,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.015566283836960793,
+ "learning_rate": 0.002498047571972777,
+ "loss": 4.815671920776367,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.012440434657037258,
+ "learning_rate": 0.002497048120005623,
+ "loss": 4.769775867462158,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01175499614328146,
+ "learning_rate": 0.0024960478768200362,
+ "loss": 4.8101348876953125,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.012699508108198643,
+ "learning_rate": 0.002495046843221896,
+ "loss": 4.827281951904297,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.012438303790986538,
+ "learning_rate": 0.002494045020017715,
+ "loss": 4.786289691925049,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.01709924265742302,
+ "learning_rate": 0.002493042408014646,
+ "loss": 4.828742027282715,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.022684535011649132,
+ "learning_rate": 0.0024920390080204733,
+ "loss": 4.822258949279785,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.011984334327280521,
+ "learning_rate": 0.00249103482084362,
+ "loss": 4.792006969451904,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.013941102661192417,
+ "learning_rate": 0.00249002984729314,
+ "loss": 4.7926177978515625,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.0145634226500988,
+ "learning_rate": 0.002489024088178724,
+ "loss": 4.819155216217041,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.013391612097620964,
+ "learning_rate": 0.0024880175443106932,
+ "loss": 4.777242660522461,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01764756441116333,
+ "learning_rate": 0.0024870102165000015,
+ "loss": 4.81627893447876,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.015397441573441029,
+ "learning_rate": 0.002486002105558235,
+ "loss": 4.869495391845703,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01225620973855257,
+ "learning_rate": 0.0024849932122976105,
+ "loss": 4.784395217895508,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.01534759160131216,
+ "learning_rate": 0.002483983537530975,
+ "loss": 4.813475608825684,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.015057779848575592,
+ "learning_rate": 0.002482973082071805,
+ "loss": 4.8281569480896,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.01243986189365387,
+ "learning_rate": 0.0024819618467342066,
+ "loss": 4.7866973876953125,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.013326527550816536,
+ "learning_rate": 0.0024809498323329136,
+ "loss": 4.750214576721191,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.013644695281982422,
+ "learning_rate": 0.002479937039683288,
+ "loss": 4.790144920349121,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01296213548630476,
+ "learning_rate": 0.002478923469601319,
+ "loss": 4.812170028686523,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.011848201975226402,
+ "learning_rate": 0.0024779091229036204,
+ "loss": 4.798855781555176,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.012765968218445778,
+ "learning_rate": 0.0024768940004074335,
+ "loss": 4.8272294998168945,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.011700666509568691,
+ "learning_rate": 0.002475878102930625,
+ "loss": 4.762265205383301,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.01018498931080103,
+ "learning_rate": 0.0024748614312916843,
+ "loss": 4.796628952026367,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.010241681709885597,
+ "learning_rate": 0.002473843986309725,
+ "loss": 4.829319000244141,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.009332080371677876,
+ "learning_rate": 0.0024728257688044855,
+ "loss": 4.783773899078369,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.012860726565122604,
+ "learning_rate": 0.002471806779596324,
+ "loss": 4.802700519561768,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.011156697757542133,
+ "learning_rate": 0.0024707870195062225,
+ "loss": 4.808038234710693,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.010329720564186573,
+ "learning_rate": 0.002469766489355783,
+ "loss": 4.761645317077637,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.011093895882368088,
+ "learning_rate": 0.0024687451899672278,
+ "loss": 4.795220851898193,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.010740782134234905,
+ "learning_rate": 0.0024677231221634,
+ "loss": 4.79289436340332,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.012272508814930916,
+ "learning_rate": 0.0024667002867677606,
+ "loss": 4.811873435974121,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.008665879257023335,
+ "learning_rate": 0.0024656766846043905,
+ "loss": 4.744179725646973,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.00933373998850584,
+ "learning_rate": 0.0024646523164979866,
+ "loss": 4.805453777313232,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.010399337857961655,
+ "learning_rate": 0.002463627183273864,
+ "loss": 4.75568962097168,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.013210986740887165,
+ "learning_rate": 0.0024626012857579542,
+ "loss": 4.797974586486816,
+ "step": 1080
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 4.604420041998336e+17,
+ "train_batch_size": 16,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/baseline/checkpoint-1080/training_args.bin b/runs/baseline/checkpoint-1080/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7
--- /dev/null
+++ b/runs/baseline/checkpoint-1080/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97
+size 5265
diff --git a/runs/baseline/checkpoint-1440/chat_template.jinja b/runs/baseline/checkpoint-1440/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/baseline/checkpoint-1440/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/baseline/checkpoint-1440/config.json b/runs/baseline/checkpoint-1440/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/baseline/checkpoint-1440/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/baseline/checkpoint-1440/generation_config.json b/runs/baseline/checkpoint-1440/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/baseline/checkpoint-1440/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/baseline/checkpoint-1440/model.safetensors b/runs/baseline/checkpoint-1440/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..6365dbd1359578638aa2e7a4cd5e47df42b85c11
--- /dev/null
+++ b/runs/baseline/checkpoint-1440/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2d590f62abdc2e7d7be25aacb35cb7e0eb2e858370679a61e7caec37105ed0d5
+size 583356232
diff --git a/runs/baseline/checkpoint-1440/optimizer.pt b/runs/baseline/checkpoint-1440/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b80ec4b8098716014384f3d375b6e2f58c238f9b
--- /dev/null
+++ b/runs/baseline/checkpoint-1440/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6bf4afa93e6537e6e0f119f6af49e6a223f95a084b81e99ed26287e4467cd6a7
+size 1166825803
diff --git a/runs/baseline/checkpoint-1440/rng_state_0.pth b/runs/baseline/checkpoint-1440/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3c396253e5ad58aa3254e1cb9364198815e3ba61
--- /dev/null
+++ b/runs/baseline/checkpoint-1440/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ad41999a3021e51cc92b222472cf4756a254132cabf58c4c2e1ee206c7366885
+size 14917
diff --git a/runs/baseline/checkpoint-1440/rng_state_1.pth b/runs/baseline/checkpoint-1440/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..08b5f24cb1ec06dc192e471442f86bfe5e398030
--- /dev/null
+++ b/runs/baseline/checkpoint-1440/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:29848a162fabe62f2066ebb397d9dc38254fb7b578fd1c8908975df46738482d
+size 14917
diff --git a/runs/baseline/checkpoint-1440/scheduler.pt b/runs/baseline/checkpoint-1440/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..032c2c72b5314d57cafe9ae23600e7369414b19b
--- /dev/null
+++ b/runs/baseline/checkpoint-1440/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:cc198898504ac73bd5b7b2d8d7dbdfa658d335ae2db41ad71022be987a114f4e
+size 1465
diff --git a/runs/baseline/checkpoint-1440/tokenizer.json b/runs/baseline/checkpoint-1440/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/baseline/checkpoint-1440/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/baseline/checkpoint-1440/tokenizer_config.json b/runs/baseline/checkpoint-1440/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/baseline/checkpoint-1440/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/baseline/checkpoint-1440/trainer_state.json b/runs/baseline/checkpoint-1440/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..1f5b852ec31bac36850963888bdb68810449cb10
--- /dev/null
+++ b/runs/baseline/checkpoint-1440/trainer_state.json
@@ -0,0 +1,10266 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 20.0,
+ "eval_steps": 500,
+ "global_step": 1440,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.12165174633264542,
+ "learning_rate": 0.0,
+ "loss": 12.027063369750977,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.12109383195638657,
+ "learning_rate": 3e-05,
+ "loss": 12.028635025024414,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.12690064311027527,
+ "learning_rate": 6e-05,
+ "loss": 11.894524574279785,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.1537286937236786,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.672262191772461,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.11778244376182556,
+ "learning_rate": 0.00012,
+ "loss": 11.433671951293945,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.10874935984611511,
+ "learning_rate": 0.00015000000000000001,
+ "loss": 11.270309448242188,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.10641921311616898,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 11.155481338500977,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.1048506498336792,
+ "learning_rate": 0.00021000000000000004,
+ "loss": 11.066675186157227,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.10546834021806717,
+ "learning_rate": 0.00024,
+ "loss": 10.952295303344727,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.10614268481731415,
+ "learning_rate": 0.00027,
+ "loss": 10.811386108398438,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10529010742902756,
+ "learning_rate": 0.00030000000000000003,
+ "loss": 10.657712936401367,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.1051444336771965,
+ "learning_rate": 0.00033,
+ "loss": 10.47591781616211,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10420846194028854,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 10.280141830444336,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10323625057935715,
+ "learning_rate": 0.00039000000000000005,
+ "loss": 10.076366424560547,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10255401581525803,
+ "learning_rate": 0.00042000000000000007,
+ "loss": 9.856939315795898,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10151027888059616,
+ "learning_rate": 0.00045,
+ "loss": 9.632134437561035,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.1006326898932457,
+ "learning_rate": 0.00048,
+ "loss": 9.401637077331543,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.09820562601089478,
+ "learning_rate": 0.00051,
+ "loss": 9.177400588989258,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.0956626906991005,
+ "learning_rate": 0.00054,
+ "loss": 8.95518684387207,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.0922091156244278,
+ "learning_rate": 0.00057,
+ "loss": 8.75066089630127,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.08593932539224625,
+ "learning_rate": 0.0006000000000000001,
+ "loss": 8.557976722717285,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.08014766871929169,
+ "learning_rate": 0.00063,
+ "loss": 8.360604286193848,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.07381244748830795,
+ "learning_rate": 0.00066,
+ "loss": 8.210375785827637,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.06455052644014359,
+ "learning_rate": 0.0006900000000000001,
+ "loss": 8.07168960571289,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.054115474224090576,
+ "learning_rate": 0.0007199999999999999,
+ "loss": 7.960069179534912,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.043575264513492584,
+ "learning_rate": 0.00075,
+ "loss": 7.850685119628906,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.02976679615676403,
+ "learning_rate": 0.0007800000000000001,
+ "loss": 7.824377059936523,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.020552165806293488,
+ "learning_rate": 0.0008100000000000001,
+ "loss": 7.7695417404174805,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.018530743196606636,
+ "learning_rate": 0.0008400000000000001,
+ "loss": 7.782247543334961,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.02716820128262043,
+ "learning_rate": 0.00087,
+ "loss": 7.754945278167725,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.03638409823179245,
+ "learning_rate": 0.0009,
+ "loss": 7.799381256103516,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.041579652577638626,
+ "learning_rate": 0.00093,
+ "loss": 7.776629447937012,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.0304960235953331,
+ "learning_rate": 0.00096,
+ "loss": 7.749439239501953,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.053728822618722916,
+ "learning_rate": 0.00099,
+ "loss": 7.745059490203857,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.036322493106126785,
+ "learning_rate": 0.00102,
+ "loss": 7.752074718475342,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.03232959285378456,
+ "learning_rate": 0.00105,
+ "loss": 7.704257011413574,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.03246184438467026,
+ "learning_rate": 0.00108,
+ "loss": 7.667045593261719,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.036655545234680176,
+ "learning_rate": 0.00111,
+ "loss": 7.623445987701416,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.021042725071310997,
+ "learning_rate": 0.00114,
+ "loss": 7.582821846008301,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.022982975468039513,
+ "learning_rate": 0.00117,
+ "loss": 7.58036994934082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.024781864136457443,
+ "learning_rate": 0.0012000000000000001,
+ "loss": 7.5903000831604,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.03667089343070984,
+ "learning_rate": 0.00123,
+ "loss": 7.588233947753906,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.04083017632365227,
+ "learning_rate": 0.00126,
+ "loss": 7.548343658447266,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.02585385926067829,
+ "learning_rate": 0.00129,
+ "loss": 7.604833602905273,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.018793931230902672,
+ "learning_rate": 0.00132,
+ "loss": 7.56705379486084,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.025538936257362366,
+ "learning_rate": 0.00135,
+ "loss": 7.562107086181641,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.02819318138062954,
+ "learning_rate": 0.0013800000000000002,
+ "loss": 7.517213821411133,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.016239255666732788,
+ "learning_rate": 0.00141,
+ "loss": 7.4780755043029785,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.024470701813697815,
+ "learning_rate": 0.0014399999999999999,
+ "loss": 7.505887985229492,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.015499847941100597,
+ "learning_rate": 0.00147,
+ "loss": 7.484469890594482,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.017785200849175453,
+ "learning_rate": 0.0015,
+ "loss": 7.481822967529297,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.015952805057168007,
+ "learning_rate": 0.0015300000000000001,
+ "loss": 7.484804153442383,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.015771260485053062,
+ "learning_rate": 0.0015600000000000002,
+ "loss": 7.4495439529418945,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.024370530620217323,
+ "learning_rate": 0.00159,
+ "loss": 7.471208572387695,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.018173737451434135,
+ "learning_rate": 0.0016200000000000001,
+ "loss": 7.438447952270508,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.020564094185829163,
+ "learning_rate": 0.0016500000000000002,
+ "loss": 7.437922954559326,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.016154944896697998,
+ "learning_rate": 0.0016800000000000003,
+ "loss": 7.429378032684326,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.016817007213830948,
+ "learning_rate": 0.00171,
+ "loss": 7.406428337097168,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.03460947796702385,
+ "learning_rate": 0.00174,
+ "loss": 7.393331527709961,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.15410488843917847,
+ "learning_rate": 0.0017699999999999999,
+ "loss": 7.645856857299805,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.07443148642778397,
+ "learning_rate": 0.0018,
+ "loss": 7.549508094787598,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.0507572740316391,
+ "learning_rate": 0.00183,
+ "loss": 7.517921447753906,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.09682746231555939,
+ "learning_rate": 0.00186,
+ "loss": 7.543978691101074,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.2571517527103424,
+ "learning_rate": 0.00189,
+ "loss": 7.768949508666992,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.0317784883081913,
+ "learning_rate": 0.00192,
+ "loss": 7.509063720703125,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.0353970006108284,
+ "learning_rate": 0.0019500000000000001,
+ "loss": 7.522322654724121,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.03663994371891022,
+ "learning_rate": 0.00198,
+ "loss": 7.504723072052002,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.025481393560767174,
+ "learning_rate": 0.00201,
+ "loss": 7.452597141265869,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.03099915198981762,
+ "learning_rate": 0.00204,
+ "loss": 7.455441474914551,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.022921698167920113,
+ "learning_rate": 0.00207,
+ "loss": 7.4454827308654785,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.032388731837272644,
+ "learning_rate": 0.0021,
+ "loss": 7.467103958129883,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.022428609430789948,
+ "learning_rate": 0.00213,
+ "loss": 7.467540264129639,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.469719409942627,
+ "eval_runtime": 50.9714,
+ "eval_samples_per_second": 47.909,
+ "eval_steps_per_second": 1.511,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.017480911687016487,
+ "learning_rate": 0.00216,
+ "loss": 7.455810546875,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.016675131395459175,
+ "learning_rate": 0.00219,
+ "loss": 7.474848747253418,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.017450228333473206,
+ "learning_rate": 0.00222,
+ "loss": 7.442611217498779,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.016823675483465195,
+ "learning_rate": 0.0022500000000000003,
+ "loss": 7.454628944396973,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.016717543825507164,
+ "learning_rate": 0.00228,
+ "loss": 7.425317764282227,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.01644429750740528,
+ "learning_rate": 0.00231,
+ "loss": 7.4102864265441895,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.013104400597512722,
+ "learning_rate": 0.00234,
+ "loss": 7.4559173583984375,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.01175684668123722,
+ "learning_rate": 0.00237,
+ "loss": 7.438898086547852,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.013736684806644917,
+ "learning_rate": 0.0024000000000000002,
+ "loss": 7.449005603790283,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01176756713539362,
+ "learning_rate": 0.0024300000000000003,
+ "loss": 7.428421974182129,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.011860918253660202,
+ "learning_rate": 0.00246,
+ "loss": 7.411491394042969,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.010093477554619312,
+ "learning_rate": 0.00249,
+ "loss": 7.39311408996582,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.009453288279473782,
+ "learning_rate": 0.00252,
+ "loss": 7.387656211853027,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.010311692021787167,
+ "learning_rate": 0.00255,
+ "loss": 7.401076316833496,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.010053195990622044,
+ "learning_rate": 0.00258,
+ "loss": 7.39655876159668,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.009758037514984608,
+ "learning_rate": 0.00261,
+ "loss": 7.423122882843018,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008394946344196796,
+ "learning_rate": 0.00264,
+ "loss": 7.383601188659668,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.008457187563180923,
+ "learning_rate": 0.00267,
+ "loss": 7.400934219360352,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.010601334273815155,
+ "learning_rate": 0.0027,
+ "loss": 7.400241851806641,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.009009373374283314,
+ "learning_rate": 0.0027300000000000002,
+ "loss": 7.377348899841309,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.00737508200109005,
+ "learning_rate": 0.0027600000000000003,
+ "loss": 7.385881423950195,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.0075711836107075214,
+ "learning_rate": 0.0027900000000000004,
+ "loss": 7.348570823669434,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.00846901722252369,
+ "learning_rate": 0.00282,
+ "loss": 7.364287853240967,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.007160635199397802,
+ "learning_rate": 0.00285,
+ "loss": 7.3881635665893555,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.00682067358866334,
+ "learning_rate": 0.0028799999999999997,
+ "loss": 7.348663330078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.009108101949095726,
+ "learning_rate": 0.00291,
+ "loss": 7.356302261352539,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.015828927978873253,
+ "learning_rate": 0.00294,
+ "loss": 7.347576141357422,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.009683096781373024,
+ "learning_rate": 0.00297,
+ "loss": 7.323967933654785,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.010833270847797394,
+ "learning_rate": 0.003,
+ "loss": 7.3139495849609375,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.014905400574207306,
+ "learning_rate": 0.002999999401781161,
+ "loss": 7.302030086517334,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.023674853146076202,
+ "learning_rate": 0.002999997607125127,
+ "loss": 7.311811923980713,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.050460800528526306,
+ "learning_rate": 0.002999994616033342,
+ "loss": 7.41972541809082,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.02193174883723259,
+ "learning_rate": 0.002999990428508218,
+ "loss": 7.294468879699707,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.036261629313230515,
+ "learning_rate": 0.0029999850445531283,
+ "loss": 7.295568466186523,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.02734929881989956,
+ "learning_rate": 0.002999978464172409,
+ "loss": 7.309239864349365,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.017884107306599617,
+ "learning_rate": 0.002999970687371363,
+ "loss": 7.2427592277526855,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01700308918952942,
+ "learning_rate": 0.0029999617141562573,
+ "loss": 7.241730213165283,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.0209935512393713,
+ "learning_rate": 0.0029999515445343194,
+ "loss": 7.262085437774658,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.023207105696201324,
+ "learning_rate": 0.0029999401785137437,
+ "loss": 7.25450325012207,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.01907353848218918,
+ "learning_rate": 0.002999927616103687,
+ "loss": 7.23149299621582,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.016149666160345078,
+ "learning_rate": 0.0029999138573142715,
+ "loss": 7.210232734680176,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.017513928934931755,
+ "learning_rate": 0.002999898902156582,
+ "loss": 7.218451976776123,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.015002547763288021,
+ "learning_rate": 0.002999882750642667,
+ "loss": 7.169316291809082,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.014233081601560116,
+ "learning_rate": 0.0029998654027855406,
+ "loss": 7.194825172424316,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.01283260714262724,
+ "learning_rate": 0.0029998468585991783,
+ "loss": 7.15669059753418,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.014416799880564213,
+ "learning_rate": 0.0029998271180985223,
+ "loss": 7.143369674682617,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.014337058179080486,
+ "learning_rate": 0.0029998061812994757,
+ "loss": 7.127734184265137,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.016406210139393806,
+ "learning_rate": 0.002999784048218908,
+ "loss": 7.086504936218262,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.01680421642959118,
+ "learning_rate": 0.002999760718874652,
+ "loss": 7.089420318603516,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.010665548034012318,
+ "learning_rate": 0.0029997361932855016,
+ "loss": 7.102463722229004,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013479137793183327,
+ "learning_rate": 0.002999710471471218,
+ "loss": 7.079098701477051,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.012408486567437649,
+ "learning_rate": 0.0029996835534525246,
+ "loss": 7.0790114402771,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.012035196647047997,
+ "learning_rate": 0.0029996554392511094,
+ "loss": 7.058756351470947,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.00899985246360302,
+ "learning_rate": 0.0029996261288896225,
+ "loss": 7.0739240646362305,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.01072906143963337,
+ "learning_rate": 0.002999595622391679,
+ "loss": 7.085537910461426,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.01047342736274004,
+ "learning_rate": 0.0029995639197818577,
+ "loss": 7.060049057006836,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.02342618815600872,
+ "learning_rate": 0.0029995310210857006,
+ "loss": 7.039482116699219,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.011256859637796879,
+ "learning_rate": 0.0029994969263297134,
+ "loss": 7.040412902832031,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.008407019078731537,
+ "learning_rate": 0.0029994616355413666,
+ "loss": 7.0364274978637695,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.011006148532032967,
+ "learning_rate": 0.0029994251487490916,
+ "loss": 6.991468906402588,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.009103255346417427,
+ "learning_rate": 0.0029993874659822875,
+ "loss": 7.001059532165527,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.007875788025557995,
+ "learning_rate": 0.0029993485872713125,
+ "loss": 6.9974260330200195,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.007872611284255981,
+ "learning_rate": 0.0029993085126474913,
+ "loss": 6.980216979980469,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.010071957483887672,
+ "learning_rate": 0.0029992672421431115,
+ "loss": 6.97799015045166,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.01209618803113699,
+ "learning_rate": 0.0029992247757914237,
+ "loss": 6.97202205657959,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.014884858392179012,
+ "learning_rate": 0.0029991811136266425,
+ "loss": 6.976570129394531,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.014953463338315487,
+ "learning_rate": 0.0029991362556839453,
+ "loss": 6.971050262451172,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.008371063508093357,
+ "learning_rate": 0.002999090201999474,
+ "loss": 6.955556869506836,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.011926881037652493,
+ "learning_rate": 0.002999042952610332,
+ "loss": 6.943947792053223,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.013422893360257149,
+ "learning_rate": 0.002998994507554589,
+ "loss": 6.910022735595703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.009672224521636963,
+ "learning_rate": 0.002998944866871275,
+ "loss": 6.8994855880737305,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.012590581551194191,
+ "learning_rate": 0.0029988940306003853,
+ "loss": 6.88709020614624,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.945146083831787,
+ "eval_runtime": 51.5801,
+ "eval_samples_per_second": 47.344,
+ "eval_steps_per_second": 1.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.016554836183786392,
+ "learning_rate": 0.002998841998782877,
+ "loss": 6.916609764099121,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.011085660196840763,
+ "learning_rate": 0.0029987887714606723,
+ "loss": 6.884910583496094,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.008596168830990791,
+ "learning_rate": 0.0029987343486766547,
+ "loss": 6.900066375732422,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.014487302862107754,
+ "learning_rate": 0.0029986787304746715,
+ "loss": 6.8693342208862305,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.015636997297406197,
+ "learning_rate": 0.002998621916899534,
+ "loss": 6.903670787811279,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.013275462202727795,
+ "learning_rate": 0.0029985639079970153,
+ "loss": 6.889361381530762,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.009044744074344635,
+ "learning_rate": 0.002998504703813852,
+ "loss": 6.8549981117248535,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.011327696964144707,
+ "learning_rate": 0.002998444304397745,
+ "loss": 6.856151103973389,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.01310331467539072,
+ "learning_rate": 0.0029983827097973554,
+ "loss": 6.834944725036621,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.010003753937780857,
+ "learning_rate": 0.00299831992006231,
+ "loss": 6.825613975524902,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.009928115643560886,
+ "learning_rate": 0.0029982559352431967,
+ "loss": 6.828935146331787,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.012302150949835777,
+ "learning_rate": 0.0029981907553915674,
+ "loss": 6.821207046508789,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.020555168390274048,
+ "learning_rate": 0.002998124380559936,
+ "loss": 6.8184895515441895,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.011417638510465622,
+ "learning_rate": 0.0029980568108017797,
+ "loss": 6.788754463195801,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.010772978886961937,
+ "learning_rate": 0.002997988046171538,
+ "loss": 6.7861433029174805,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.012753859162330627,
+ "learning_rate": 0.002997918086724614,
+ "loss": 6.788776397705078,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.011327114887535572,
+ "learning_rate": 0.0029978469325173717,
+ "loss": 6.779160022735596,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01235753484070301,
+ "learning_rate": 0.002997774583607139,
+ "loss": 6.799554824829102,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.01417156495153904,
+ "learning_rate": 0.0029977010400522074,
+ "loss": 6.77957820892334,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.021321553736925125,
+ "learning_rate": 0.002997626301911828,
+ "loss": 6.785632133483887,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.03614472970366478,
+ "learning_rate": 0.002997550369246217,
+ "loss": 6.836414337158203,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03934052586555481,
+ "learning_rate": 0.002997473242116551,
+ "loss": 6.854659557342529,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.016240060329437256,
+ "learning_rate": 0.0029973949205849707,
+ "loss": 6.789690971374512,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.031164905056357384,
+ "learning_rate": 0.0029973154047145782,
+ "loss": 6.8363776206970215,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.013051844201982021,
+ "learning_rate": 0.002997234694569438,
+ "loss": 6.803576469421387,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.022611448541283607,
+ "learning_rate": 0.002997152790214577,
+ "loss": 6.762107849121094,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.017812907695770264,
+ "learning_rate": 0.0029970696917159833,
+ "loss": 6.776218414306641,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.015453332103788853,
+ "learning_rate": 0.0029969853991406088,
+ "loss": 6.742618083953857,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.020630091428756714,
+ "learning_rate": 0.0029968999125563655,
+ "loss": 6.765475749969482,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.014672341756522655,
+ "learning_rate": 0.0029968132320321297,
+ "loss": 6.754009246826172,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.016540220007300377,
+ "learning_rate": 0.0029967253576377376,
+ "loss": 6.734752655029297,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.013495649211108685,
+ "learning_rate": 0.0029966362894439874,
+ "loss": 6.752676010131836,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015936989337205887,
+ "learning_rate": 0.0029965460275226405,
+ "loss": 6.725708961486816,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010431243106722832,
+ "learning_rate": 0.0029964545719464194,
+ "loss": 6.698376655578613,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01669321395456791,
+ "learning_rate": 0.0029963619227890083,
+ "loss": 6.711770534515381,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.013308997265994549,
+ "learning_rate": 0.0029962680801250514,
+ "loss": 6.705286979675293,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011648476123809814,
+ "learning_rate": 0.002996173044030159,
+ "loss": 6.690052509307861,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.03942161798477173,
+ "learning_rate": 0.0029960768145808965,
+ "loss": 6.711885452270508,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.01571274921298027,
+ "learning_rate": 0.002995979391854797,
+ "loss": 6.715847969055176,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.00829714722931385,
+ "learning_rate": 0.0029958807759303515,
+ "loss": 6.685308933258057,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.01374727487564087,
+ "learning_rate": 0.0029957809668870114,
+ "loss": 6.695744514465332,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.010857059620320797,
+ "learning_rate": 0.0029956799648051934,
+ "loss": 6.658294677734375,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.008897246792912483,
+ "learning_rate": 0.002995577769766272,
+ "loss": 6.642439842224121,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015880772843956947,
+ "learning_rate": 0.002995474381852584,
+ "loss": 6.674907684326172,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.021623898297548294,
+ "learning_rate": 0.0029953698011474265,
+ "loss": 6.66236686706543,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.03084627166390419,
+ "learning_rate": 0.0029952640277350596,
+ "loss": 6.666715621948242,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.040561504662036896,
+ "learning_rate": 0.002995157061700702,
+ "loss": 6.649062156677246,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.031126515939831734,
+ "learning_rate": 0.0029950489031305345,
+ "loss": 6.640542030334473,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.00872360821813345,
+ "learning_rate": 0.0029949395521116987,
+ "loss": 6.635221481323242,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.023432321846485138,
+ "learning_rate": 0.0029948290087322966,
+ "loss": 6.642196178436279,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.021096013486385345,
+ "learning_rate": 0.0029947172730813913,
+ "loss": 6.643655300140381,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.014494858682155609,
+ "learning_rate": 0.002994604345249006,
+ "loss": 6.612720966339111,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.05225599184632301,
+ "learning_rate": 0.0029944902253261248,
+ "loss": 6.60659646987915,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.014311269856989384,
+ "learning_rate": 0.0029943749134046915,
+ "loss": 6.54373025894165,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.010777472518384457,
+ "learning_rate": 0.0029942584095776115,
+ "loss": 6.599837303161621,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.012498079799115658,
+ "learning_rate": 0.002994140713938749,
+ "loss": 6.589657783508301,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.015775132924318314,
+ "learning_rate": 0.00299402182658293,
+ "loss": 6.577511787414551,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.017978230491280556,
+ "learning_rate": 0.0029939017476059396,
+ "loss": 6.58717155456543,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.026807745918631554,
+ "learning_rate": 0.0029937804771045236,
+ "loss": 6.578145503997803,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.04853282868862152,
+ "learning_rate": 0.0029936580151763872,
+ "loss": 6.609755516052246,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.1057916209101677,
+ "learning_rate": 0.002993534361920195,
+ "loss": 6.870367050170898,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.04925911873579025,
+ "learning_rate": 0.002993409517435574,
+ "loss": 6.871556282043457,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.05288556218147278,
+ "learning_rate": 0.0029932834818231077,
+ "loss": 6.808315753936768,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.03831710293889046,
+ "learning_rate": 0.0029931562551843405,
+ "loss": 6.827617645263672,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.03665117174386978,
+ "learning_rate": 0.002993027837621778,
+ "loss": 6.767284870147705,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.026216736063361168,
+ "learning_rate": 0.0029928982292388826,
+ "loss": 6.733659744262695,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.024796854704618454,
+ "learning_rate": 0.002992767430140078,
+ "loss": 6.759577751159668,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.02410992607474327,
+ "learning_rate": 0.002992635440430747,
+ "loss": 6.701825141906738,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.019924134016036987,
+ "learning_rate": 0.0029925022602172308,
+ "loss": 6.677121162414551,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.022516867145895958,
+ "learning_rate": 0.00299236788960683,
+ "loss": 6.668622970581055,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.025068556889891624,
+ "learning_rate": 0.002992232328707806,
+ "loss": 6.67718505859375,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.02574208751320839,
+ "learning_rate": 0.002992095577629376,
+ "loss": 6.679594039916992,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.657555103302002,
+ "eval_runtime": 51.0199,
+ "eval_samples_per_second": 47.864,
+ "eval_steps_per_second": 1.509,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.017298126593232155,
+ "learning_rate": 0.0029919576364817184,
+ "loss": 6.620774269104004,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.018840325996279716,
+ "learning_rate": 0.002991818505375971,
+ "loss": 6.613316535949707,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.020813113078475,
+ "learning_rate": 0.002991678184424228,
+ "loss": 6.621555328369141,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.026581479236483574,
+ "learning_rate": 0.0029915366737395437,
+ "loss": 6.613471031188965,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.02084030583500862,
+ "learning_rate": 0.0029913939734359306,
+ "loss": 6.559704780578613,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.01327502354979515,
+ "learning_rate": 0.00299125008362836,
+ "loss": 6.564910888671875,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.015091734007000923,
+ "learning_rate": 0.0029911050044327614,
+ "loss": 6.571262836456299,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.015463982708752155,
+ "learning_rate": 0.002990958735966023,
+ "loss": 6.565750598907471,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.013636826537549496,
+ "learning_rate": 0.0029908112783459894,
+ "loss": 6.57981014251709,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.010613273829221725,
+ "learning_rate": 0.0029906626316914655,
+ "loss": 6.538505554199219,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.01804237812757492,
+ "learning_rate": 0.0029905127961222127,
+ "loss": 6.5282511711120605,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.015131509862840176,
+ "learning_rate": 0.0029903617717589506,
+ "loss": 6.531335830688477,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.010518206283450127,
+ "learning_rate": 0.002990209558723358,
+ "loss": 6.521886825561523,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.011360022239387035,
+ "learning_rate": 0.002990056157138069,
+ "loss": 6.528301239013672,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.011556181125342846,
+ "learning_rate": 0.0029899015671266773,
+ "loss": 6.497917652130127,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.01365791168063879,
+ "learning_rate": 0.0029897457888137325,
+ "loss": 6.52849006652832,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.01091043371707201,
+ "learning_rate": 0.0029895888223247426,
+ "loss": 6.492197036743164,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.008484847843647003,
+ "learning_rate": 0.002989430667786173,
+ "loss": 6.497377395629883,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.00836879201233387,
+ "learning_rate": 0.0029892713253254466,
+ "loss": 6.495189666748047,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.009781497530639172,
+ "learning_rate": 0.0029891107950709406,
+ "loss": 6.454085826873779,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.009603648446500301,
+ "learning_rate": 0.0029889490771519934,
+ "loss": 6.499475479125977,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.007991855964064598,
+ "learning_rate": 0.0029887861716988975,
+ "loss": 6.460785865783691,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.008447187952697277,
+ "learning_rate": 0.0029886220788429025,
+ "loss": 6.504582405090332,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.010945498012006283,
+ "learning_rate": 0.0029884567987162156,
+ "loss": 6.49589729309082,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.011096215806901455,
+ "learning_rate": 0.0029882903314519997,
+ "loss": 6.460657119750977,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.013094739057123661,
+ "learning_rate": 0.0029881226771843747,
+ "loss": 6.462450981140137,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.015423037111759186,
+ "learning_rate": 0.0029879538360484165,
+ "loss": 6.45565128326416,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.0182222630828619,
+ "learning_rate": 0.002987783808180157,
+ "loss": 6.452227592468262,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.01981227844953537,
+ "learning_rate": 0.002987612593716585,
+ "loss": 6.477701663970947,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.026733700186014175,
+ "learning_rate": 0.002987440192795645,
+ "loss": 6.452212333679199,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.03504155948758125,
+ "learning_rate": 0.0029872666055562372,
+ "loss": 6.509665489196777,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.027970125898718834,
+ "learning_rate": 0.002987091832138217,
+ "loss": 6.487089157104492,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.013714591972529888,
+ "learning_rate": 0.0029869158726823974,
+ "loss": 6.469569206237793,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.016472695395350456,
+ "learning_rate": 0.0029867387273305453,
+ "loss": 6.461688995361328,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.018457358703017235,
+ "learning_rate": 0.0029865603962253833,
+ "loss": 6.490985870361328,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.011045100167393684,
+ "learning_rate": 0.0029863808795105896,
+ "loss": 6.403701305389404,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.014126897789537907,
+ "learning_rate": 0.0029862001773307972,
+ "loss": 6.379804611206055,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.014015335589647293,
+ "learning_rate": 0.0029860182898315954,
+ "loss": 6.404377460479736,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.010032658465206623,
+ "learning_rate": 0.002985835217159527,
+ "loss": 6.40291690826416,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.015469838865101337,
+ "learning_rate": 0.0029856509594620912,
+ "loss": 6.4414520263671875,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.011106910184025764,
+ "learning_rate": 0.0029854655168877397,
+ "loss": 6.396444320678711,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.01262320950627327,
+ "learning_rate": 0.0029852788895858816,
+ "loss": 6.385222434997559,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.012783976271748543,
+ "learning_rate": 0.002985091077706878,
+ "loss": 6.3818135261535645,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.01224067248404026,
+ "learning_rate": 0.0029849020814020468,
+ "loss": 6.37886905670166,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.010540086776018143,
+ "learning_rate": 0.0029847119008236576,
+ "loss": 6.37799072265625,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.012479913420975208,
+ "learning_rate": 0.0029845205361249366,
+ "loss": 6.365094184875488,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.010080077685415745,
+ "learning_rate": 0.0029843279874600622,
+ "loss": 6.371852874755859,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.013266298919916153,
+ "learning_rate": 0.002984134254984168,
+ "loss": 6.380039215087891,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.01014771405607462,
+ "learning_rate": 0.002983939338853339,
+ "loss": 6.370337963104248,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.008164579048752785,
+ "learning_rate": 0.0029837432392246185,
+ "loss": 6.331899642944336,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.01020442508161068,
+ "learning_rate": 0.002983545956255998,
+ "loss": 6.35077428817749,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.009898579679429531,
+ "learning_rate": 0.002983347490106426,
+ "loss": 6.371985912322998,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.009668919257819653,
+ "learning_rate": 0.002983147840935802,
+ "loss": 6.356776237487793,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.012560366652905941,
+ "learning_rate": 0.0029829470089049815,
+ "loss": 6.357911109924316,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.018435249105095863,
+ "learning_rate": 0.0029827449941757698,
+ "loss": 6.366821765899658,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.029046855866909027,
+ "learning_rate": 0.002982541796910927,
+ "loss": 6.367595672607422,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.043010637164115906,
+ "learning_rate": 0.002982337417274165,
+ "loss": 6.413588523864746,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.045470818877220154,
+ "learning_rate": 0.0029821318554301497,
+ "loss": 6.435015678405762,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.02148726023733616,
+ "learning_rate": 0.002981925111544498,
+ "loss": 6.330894470214844,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.026808688417077065,
+ "learning_rate": 0.00298171718578378,
+ "loss": 6.384839057922363,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.018495213240385056,
+ "learning_rate": 0.002981508078315517,
+ "loss": 6.36277961730957,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.025393597781658173,
+ "learning_rate": 0.0029812977893081834,
+ "loss": 6.3702239990234375,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020856289193034172,
+ "learning_rate": 0.002981086318931206,
+ "loss": 6.354547500610352,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.017827816307544708,
+ "learning_rate": 0.002980873667354962,
+ "loss": 6.339264869689941,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.02639714442193508,
+ "learning_rate": 0.002980659834750781,
+ "loss": 6.3254194259643555,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.015238096937537193,
+ "learning_rate": 0.002980444821290944,
+ "loss": 6.365842819213867,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.012863729149103165,
+ "learning_rate": 0.002980228627148684,
+ "loss": 6.356776714324951,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.014739899896085262,
+ "learning_rate": 0.0029800112524981845,
+ "loss": 6.332481384277344,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.013582171872258186,
+ "learning_rate": 0.0029797926975145797,
+ "loss": 6.311968803405762,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.012704637832939625,
+ "learning_rate": 0.0029795729623739555,
+ "loss": 6.321654319763184,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.012550560757517815,
+ "learning_rate": 0.0029793520472533495,
+ "loss": 6.298409461975098,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.011170165613293648,
+ "learning_rate": 0.002979129952330748,
+ "loss": 6.294830799102783,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.326757907867432,
+ "eval_runtime": 51.6566,
+ "eval_samples_per_second": 47.274,
+ "eval_steps_per_second": 1.491,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.010947446338832378,
+ "learning_rate": 0.0029789066777850897,
+ "loss": 6.286581039428711,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.010097995400428772,
+ "learning_rate": 0.002978682223796262,
+ "loss": 6.288301467895508,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.01224475260823965,
+ "learning_rate": 0.002978456590545104,
+ "loss": 6.264986991882324,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.009734537452459335,
+ "learning_rate": 0.0029782297782134045,
+ "loss": 6.304243087768555,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.01136956550180912,
+ "learning_rate": 0.0029780017869839013,
+ "loss": 6.244938373565674,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.00935293547809124,
+ "learning_rate": 0.002977772617040284,
+ "loss": 6.251852035522461,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.011118784546852112,
+ "learning_rate": 0.0029775422685671894,
+ "loss": 6.256633758544922,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.010837923735380173,
+ "learning_rate": 0.0029773107417502055,
+ "loss": 6.263411045074463,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.011920305900275707,
+ "learning_rate": 0.0029770780367758703,
+ "loss": 6.239311218261719,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.014117483049631119,
+ "learning_rate": 0.0029768441538316685,
+ "loss": 6.253314018249512,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.022502148523926735,
+ "learning_rate": 0.002976609093106036,
+ "loss": 6.246156692504883,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.03348720073699951,
+ "learning_rate": 0.002976372854788357,
+ "loss": 6.255914688110352,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.03427121415734291,
+ "learning_rate": 0.002976135439068965,
+ "loss": 6.289946556091309,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.02407355234026909,
+ "learning_rate": 0.002975896846139141,
+ "loss": 6.288534164428711,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.017861567437648773,
+ "learning_rate": 0.0029756570761911147,
+ "loss": 6.258049964904785,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.019980227574706078,
+ "learning_rate": 0.0029754161294180646,
+ "loss": 6.250895023345947,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.01443127728998661,
+ "learning_rate": 0.0029751740060141177,
+ "loss": 6.212200164794922,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.014975055120885372,
+ "learning_rate": 0.0029749307061743475,
+ "loss": 6.230568885803223,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01408296450972557,
+ "learning_rate": 0.0029746862300947773,
+ "loss": 6.243950366973877,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.011649327352643013,
+ "learning_rate": 0.0029744405779723766,
+ "loss": 6.197744846343994,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.013392188586294651,
+ "learning_rate": 0.002974193750005063,
+ "loss": 6.207277297973633,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.010902997106313705,
+ "learning_rate": 0.002973945746391701,
+ "loss": 6.166373252868652,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.012915392406284809,
+ "learning_rate": 0.0029736965673321032,
+ "loss": 6.186316013336182,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.010772830806672573,
+ "learning_rate": 0.0029734462130270273,
+ "loss": 6.172578811645508,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.010274911299347878,
+ "learning_rate": 0.002973194683678181,
+ "loss": 6.206851005554199,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.013210494071245193,
+ "learning_rate": 0.0029729419794882164,
+ "loss": 6.21115779876709,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.013372313231229782,
+ "learning_rate": 0.0029726881006607326,
+ "loss": 6.18442440032959,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.008105908520519733,
+ "learning_rate": 0.0029724330474002743,
+ "loss": 6.204354286193848,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.014167108573019505,
+ "learning_rate": 0.0029721768199123346,
+ "loss": 6.180662155151367,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.008339803665876389,
+ "learning_rate": 0.002971919418403351,
+ "loss": 6.150079727172852,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.015939386561512947,
+ "learning_rate": 0.0029716608430807073,
+ "loss": 6.175203323364258,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.010777109302580357,
+ "learning_rate": 0.0029714010941527327,
+ "loss": 6.166106224060059,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.011274735443294048,
+ "learning_rate": 0.002971140171828702,
+ "loss": 6.163111686706543,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013626808300614357,
+ "learning_rate": 0.002970878076318836,
+ "loss": 6.140850067138672,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.016907762736082077,
+ "learning_rate": 0.0029706148078343,
+ "loss": 6.192012786865234,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.018817441537976265,
+ "learning_rate": 0.002970350366587205,
+ "loss": 6.162238121032715,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.02829768694937229,
+ "learning_rate": 0.0029700847527906068,
+ "loss": 6.159066200256348,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.02034052088856697,
+ "learning_rate": 0.002969817966658505,
+ "loss": 6.164158344268799,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.01816369965672493,
+ "learning_rate": 0.0029695500084058447,
+ "loss": 6.132316589355469,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.013685261830687523,
+ "learning_rate": 0.0029692808782485146,
+ "loss": 6.152181625366211,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.009791149757802486,
+ "learning_rate": 0.0029690105764033485,
+ "loss": 6.138544082641602,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.014827069826424122,
+ "learning_rate": 0.002968739103088124,
+ "loss": 6.135766983032227,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.01391249243170023,
+ "learning_rate": 0.0029684664585215618,
+ "loss": 6.146172523498535,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.012729899026453495,
+ "learning_rate": 0.002968192642923326,
+ "loss": 6.122613430023193,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.014175458811223507,
+ "learning_rate": 0.002967917656514026,
+ "loss": 6.12574577331543,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.010958840139210224,
+ "learning_rate": 0.002967641499515213,
+ "loss": 6.111088752746582,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.015380816534161568,
+ "learning_rate": 0.0029673641721493817,
+ "loss": 6.100615501403809,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.025339096784591675,
+ "learning_rate": 0.00296708567463997,
+ "loss": 6.139659881591797,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.03695995733141899,
+ "learning_rate": 0.002966806007211358,
+ "loss": 6.160996437072754,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.0408046580851078,
+ "learning_rate": 0.0029665251700888686,
+ "loss": 6.173980712890625,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.025022024288773537,
+ "learning_rate": 0.0029662431634987674,
+ "loss": 6.145794868469238,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.027254154905676842,
+ "learning_rate": 0.002965959987668262,
+ "loss": 6.16928243637085,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.06191903352737427,
+ "learning_rate": 0.0029656756428255017,
+ "loss": 6.138525009155273,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0348956398665905,
+ "learning_rate": 0.002965390129199579,
+ "loss": 6.185477256774902,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.020859677344560623,
+ "learning_rate": 0.002965103447020526,
+ "loss": 6.148846626281738,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.022810814902186394,
+ "learning_rate": 0.0029648155965193174,
+ "loss": 6.1701178550720215,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.017981817945837975,
+ "learning_rate": 0.00296452657792787,
+ "loss": 6.1323771476745605,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.014905866235494614,
+ "learning_rate": 0.00296423639147904,
+ "loss": 6.123426914215088,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.01463468000292778,
+ "learning_rate": 0.002963945037406626,
+ "loss": 6.146122932434082,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.021616583690047264,
+ "learning_rate": 0.002963652515945366,
+ "loss": 6.093459606170654,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.01550968922674656,
+ "learning_rate": 0.00296335882733094,
+ "loss": 6.114929676055908,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.014597428031265736,
+ "learning_rate": 0.0029630639717999675,
+ "loss": 6.113945960998535,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.018249280750751495,
+ "learning_rate": 0.002962767949590008,
+ "loss": 6.130507469177246,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.014160148799419403,
+ "learning_rate": 0.0029624707609395614,
+ "loss": 6.086946487426758,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.015563832595944405,
+ "learning_rate": 0.002962172406088067,
+ "loss": 6.083332538604736,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.012240015901625156,
+ "learning_rate": 0.002961872885275904,
+ "loss": 6.0822906494140625,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.0141651201993227,
+ "learning_rate": 0.002961572198744391,
+ "loss": 6.049562454223633,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.010004645213484764,
+ "learning_rate": 0.0029612703467357866,
+ "loss": 6.067063808441162,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.011790835298597813,
+ "learning_rate": 0.0029609673294932856,
+ "loss": 6.062193870544434,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.011088060215115547,
+ "learning_rate": 0.002960663147261025,
+ "loss": 6.070950031280518,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.010387620888650417,
+ "learning_rate": 0.002960357800284078,
+ "loss": 6.032423973083496,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.012074458412826061,
+ "learning_rate": 0.002960051288808458,
+ "loss": 6.066788673400879,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.097111701965332,
+ "eval_runtime": 51.8742,
+ "eval_samples_per_second": 47.075,
+ "eval_steps_per_second": 1.484,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.013429952785372734,
+ "learning_rate": 0.0029597436130811153,
+ "loss": 6.022172451019287,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.01764705777168274,
+ "learning_rate": 0.0029594347733499387,
+ "loss": 6.029745101928711,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.02403099089860916,
+ "learning_rate": 0.002959124769863755,
+ "loss": 6.04707145690918,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.029030559584498405,
+ "learning_rate": 0.002958813602872328,
+ "loss": 6.05479097366333,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.029980644583702087,
+ "learning_rate": 0.0029585012726263595,
+ "loss": 6.068657875061035,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.021424276754260063,
+ "learning_rate": 0.0029581877793774885,
+ "loss": 6.062008857727051,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.018076395615935326,
+ "learning_rate": 0.002957873123378291,
+ "loss": 6.023279190063477,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.017688792198896408,
+ "learning_rate": 0.002957557304882279,
+ "loss": 6.014641284942627,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.02313626930117607,
+ "learning_rate": 0.0029572403241439026,
+ "loss": 6.026420593261719,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.04045062139630318,
+ "learning_rate": 0.0029569221814185467,
+ "loss": 6.007199764251709,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.015792209655046463,
+ "learning_rate": 0.0029566028769625334,
+ "loss": 6.017209053039551,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.01990479789674282,
+ "learning_rate": 0.002956282411033121,
+ "loss": 5.982841491699219,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.01191043108701706,
+ "learning_rate": 0.002955960783888503,
+ "loss": 6.0035929679870605,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.01732831820845604,
+ "learning_rate": 0.002955637995787807,
+ "loss": 6.025006294250488,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.032948266714811325,
+ "learning_rate": 0.0029553140469910995,
+ "loss": 6.007736682891846,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.011192727833986282,
+ "learning_rate": 0.00295498893775938,
+ "loss": 6.0022687911987305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014646840281784534,
+ "learning_rate": 0.0029546626683545824,
+ "loss": 5.976479530334473,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01589796133339405,
+ "learning_rate": 0.002954335239039576,
+ "loss": 5.980949878692627,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.015129734762012959,
+ "learning_rate": 0.0029540066500781657,
+ "loss": 6.0073394775390625,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.011140518821775913,
+ "learning_rate": 0.002953676901735088,
+ "loss": 5.976587295532227,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.0109480582177639,
+ "learning_rate": 0.0029533459942760174,
+ "loss": 6.001427173614502,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.013346790336072445,
+ "learning_rate": 0.0029530139279675583,
+ "loss": 5.96919059753418,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.012474581599235535,
+ "learning_rate": 0.002952680703077251,
+ "loss": 5.970426559448242,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01140566635876894,
+ "learning_rate": 0.002952346319873569,
+ "loss": 5.989171504974365,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.0137029392644763,
+ "learning_rate": 0.0029520107786259187,
+ "loss": 6.010115623474121,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015517761930823326,
+ "learning_rate": 0.002951674079604639,
+ "loss": 6.003649711608887,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.02173629403114319,
+ "learning_rate": 0.0029513362230810035,
+ "loss": 5.967074871063232,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02683914266526699,
+ "learning_rate": 0.002950997209327216,
+ "loss": 5.994434833526611,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.025865858420729637,
+ "learning_rate": 0.002950657038616415,
+ "loss": 6.010523796081543,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02544897235929966,
+ "learning_rate": 0.002950315711222668,
+ "loss": 5.994078159332275,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.028426019474864006,
+ "learning_rate": 0.0029499732274209777,
+ "loss": 6.020925521850586,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.026174750179052353,
+ "learning_rate": 0.0029496295874872768,
+ "loss": 5.985616683959961,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.02802138403058052,
+ "learning_rate": 0.00294928479169843,
+ "loss": 5.973659038543701,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.019348938018083572,
+ "learning_rate": 0.002948938840332233,
+ "loss": 5.990137577056885,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.016294041648507118,
+ "learning_rate": 0.0029485917336674115,
+ "loss": 5.962082862854004,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.016167785972356796,
+ "learning_rate": 0.002948243471983625,
+ "loss": 5.9594573974609375,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.014119232073426247,
+ "learning_rate": 0.00294789405556146,
+ "loss": 5.894467353820801,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.013805347494781017,
+ "learning_rate": 0.002947543484682436,
+ "loss": 5.964014530181885,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.014496497809886932,
+ "learning_rate": 0.002947191759629001,
+ "loss": 5.989920616149902,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.014813800342381,
+ "learning_rate": 0.0029468388806845333,
+ "loss": 5.9669413566589355,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.01153622381389141,
+ "learning_rate": 0.0029464848481333423,
+ "loss": 5.918018341064453,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.010525977239012718,
+ "learning_rate": 0.0029461296622606647,
+ "loss": 5.892477989196777,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.011410463601350784,
+ "learning_rate": 0.002945773323352667,
+ "loss": 5.925249099731445,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.010949424467980862,
+ "learning_rate": 0.0029454158316964464,
+ "loss": 5.90645694732666,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.00962804164737463,
+ "learning_rate": 0.002945057187580026,
+ "loss": 5.917054176330566,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.011517546139657497,
+ "learning_rate": 0.0029446973912923597,
+ "loss": 5.919445514678955,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01567666605114937,
+ "learning_rate": 0.0029443364431233285,
+ "loss": 5.884111404418945,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.016781140118837357,
+ "learning_rate": 0.002943974343363742,
+ "loss": 5.9252119064331055,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.02244546078145504,
+ "learning_rate": 0.0029436110923053375,
+ "loss": 5.933143615722656,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.028186792507767677,
+ "learning_rate": 0.00294324669024078,
+ "loss": 5.9079132080078125,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.028585223481059074,
+ "learning_rate": 0.002942881137463661,
+ "loss": 5.925262451171875,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.02781485952436924,
+ "learning_rate": 0.0029425144342685003,
+ "loss": 5.963315010070801,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.021486928686499596,
+ "learning_rate": 0.002942146580950744,
+ "loss": 5.930115699768066,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.016890818253159523,
+ "learning_rate": 0.002941777577806765,
+ "loss": 5.921439170837402,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.0169126745313406,
+ "learning_rate": 0.0029414074251338618,
+ "loss": 5.928220272064209,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.014659176580607891,
+ "learning_rate": 0.002941036123230261,
+ "loss": 5.934546947479248,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.015337883494794369,
+ "learning_rate": 0.002940663672395112,
+ "loss": 5.890145301818848,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01254988368600607,
+ "learning_rate": 0.002940290072928493,
+ "loss": 5.875492095947266,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.014658362604677677,
+ "learning_rate": 0.0029399153251314066,
+ "loss": 5.901272773742676,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.017342250794172287,
+ "learning_rate": 0.002939539429305779,
+ "loss": 5.875853538513184,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.016825608909130096,
+ "learning_rate": 0.0029391623857544643,
+ "loss": 5.941972732543945,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.016716884449124336,
+ "learning_rate": 0.002938784194781238,
+ "loss": 5.912367343902588,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.021707067266106606,
+ "learning_rate": 0.002938404856690803,
+ "loss": 5.895319938659668,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.030519135296344757,
+ "learning_rate": 0.002938024371788785,
+ "loss": 5.897165298461914,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.02572188340127468,
+ "learning_rate": 0.002937642740381733,
+ "loss": 5.879164695739746,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.016286717727780342,
+ "learning_rate": 0.002937259962777122,
+ "loss": 5.841421127319336,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.020656082779169083,
+ "learning_rate": 0.002936876039283347,
+ "loss": 5.9222259521484375,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.015350459143519402,
+ "learning_rate": 0.0029364909702097298,
+ "loss": 5.885847091674805,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01708570495247841,
+ "learning_rate": 0.0029361047558665136,
+ "loss": 5.872044563293457,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.013537376187741756,
+ "learning_rate": 0.0029357173965648634,
+ "loss": 5.866434097290039,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.013208637945353985,
+ "learning_rate": 0.002935328892616868,
+ "loss": 5.8624491691589355,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.012249554507434368,
+ "learning_rate": 0.002934939244335538,
+ "loss": 5.833232879638672,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.90985107421875,
+ "eval_runtime": 51.0696,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.01382362935692072,
+ "learning_rate": 0.0029345484520348062,
+ "loss": 5.835695266723633,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.012947693467140198,
+ "learning_rate": 0.0029341565160295262,
+ "loss": 5.826190948486328,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.013653872534632683,
+ "learning_rate": 0.002933763436635475,
+ "loss": 5.831789016723633,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.027596980333328247,
+ "learning_rate": 0.0029333692141693484,
+ "loss": 5.839963436126709,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01637990213930607,
+ "learning_rate": 0.0029329738489487648,
+ "loss": 5.805248260498047,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.02169182151556015,
+ "learning_rate": 0.0029325773412922627,
+ "loss": 5.838248252868652,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.03518201410770416,
+ "learning_rate": 0.0029321796915193015,
+ "loss": 5.834336280822754,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04983877018094063,
+ "learning_rate": 0.0029317808999502603,
+ "loss": 5.879856109619141,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05085664615035057,
+ "learning_rate": 0.0029313809669064377,
+ "loss": 5.946531295776367,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.03314858302474022,
+ "learning_rate": 0.0029309798927100532,
+ "loss": 5.909746170043945,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.034118980169296265,
+ "learning_rate": 0.0029305776776842453,
+ "loss": 5.903947830200195,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.03401247411966324,
+ "learning_rate": 0.0029301743221530705,
+ "loss": 5.936162948608398,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.043827857822179794,
+ "learning_rate": 0.0029297698264415047,
+ "loss": 5.953949451446533,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.06281330436468124,
+ "learning_rate": 0.002929364190875444,
+ "loss": 5.988225936889648,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.07216024398803711,
+ "learning_rate": 0.0029289574157817014,
+ "loss": 6.187741279602051,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.055205658078193665,
+ "learning_rate": 0.002928549501488008,
+ "loss": 6.177831649780273,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.04834286496043205,
+ "learning_rate": 0.0029281404483230133,
+ "loss": 6.209604263305664,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.0447777584195137,
+ "learning_rate": 0.0029277302566162837,
+ "loss": 6.146507263183594,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.05729430541396141,
+ "learning_rate": 0.0029273189266983035,
+ "loss": 6.280686378479004,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.0410807766020298,
+ "learning_rate": 0.002926906458900473,
+ "loss": 6.246824264526367,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.044607777148485184,
+ "learning_rate": 0.0029264928535551104,
+ "loss": 6.19956111907959,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.05141043663024902,
+ "learning_rate": 0.0029260781109954515,
+ "loss": 6.118578910827637,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.029248764738440514,
+ "learning_rate": 0.0029256622315556448,
+ "loss": 6.155360221862793,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.028403814882040024,
+ "learning_rate": 0.0029252452155707582,
+ "loss": 6.097352027893066,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.025872666388750076,
+ "learning_rate": 0.002924827063376773,
+ "loss": 6.072485446929932,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.018447555601596832,
+ "learning_rate": 0.0029244077753105874,
+ "loss": 6.077045440673828,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.023166058585047722,
+ "learning_rate": 0.0029239873517100147,
+ "loss": 6.040050983428955,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.019579261541366577,
+ "learning_rate": 0.002923565792913782,
+ "loss": 6.039424896240234,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.018369073048233986,
+ "learning_rate": 0.002923143099261531,
+ "loss": 6.021180152893066,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.014188461937010288,
+ "learning_rate": 0.0029227192710938195,
+ "loss": 5.9668049812316895,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.019567398354411125,
+ "learning_rate": 0.0029222943087521182,
+ "loss": 5.989189147949219,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.014190621674060822,
+ "learning_rate": 0.0029218682125788102,
+ "loss": 5.942726135253906,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01375526562333107,
+ "learning_rate": 0.002921440982917195,
+ "loss": 5.942777633666992,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.014203300699591637,
+ "learning_rate": 0.0029210126201114836,
+ "loss": 5.927928924560547,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.014231069944798946,
+ "learning_rate": 0.0029205831245067994,
+ "loss": 5.923510551452637,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.012771515175700188,
+ "learning_rate": 0.0029201524964491796,
+ "loss": 5.908803939819336,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.013399476185441017,
+ "learning_rate": 0.0029197207362855735,
+ "loss": 5.890079021453857,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.025862423703074455,
+ "learning_rate": 0.0029192878443638427,
+ "loss": 5.919907569885254,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.011315986514091492,
+ "learning_rate": 0.0029188538210327595,
+ "loss": 5.901529788970947,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.01235947571694851,
+ "learning_rate": 0.00291841866664201,
+ "loss": 5.903595924377441,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.012175127863883972,
+ "learning_rate": 0.0029179823815421893,
+ "loss": 5.868306636810303,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01195445004850626,
+ "learning_rate": 0.0029175449660848052,
+ "loss": 5.899898529052734,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.012801481410861015,
+ "learning_rate": 0.0029171064206222745,
+ "loss": 5.890621662139893,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.011909718625247478,
+ "learning_rate": 0.002916666745507926,
+ "loss": 5.86421012878418,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01162486057728529,
+ "learning_rate": 0.0029162259410959977,
+ "loss": 5.869139671325684,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.010923407040536404,
+ "learning_rate": 0.0029157840077416374,
+ "loss": 5.869789123535156,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.011943561024963856,
+ "learning_rate": 0.002915340945800904,
+ "loss": 5.873437881469727,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.010295520536601543,
+ "learning_rate": 0.002914896755630763,
+ "loss": 5.819953918457031,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.0093373479321599,
+ "learning_rate": 0.002914451437589092,
+ "loss": 5.829481601715088,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.01182442344725132,
+ "learning_rate": 0.002914004992034674,
+ "loss": 5.855249404907227,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.00939213763922453,
+ "learning_rate": 0.0029135574193272034,
+ "loss": 5.83876371383667,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.011648556217551231,
+ "learning_rate": 0.0029131087198272813,
+ "loss": 5.824906826019287,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.011227169074118137,
+ "learning_rate": 0.002912658893896416,
+ "loss": 5.847869873046875,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.011084851808845997,
+ "learning_rate": 0.002912207941897025,
+ "loss": 5.839183807373047,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.009923825040459633,
+ "learning_rate": 0.002911755864192432,
+ "loss": 5.818821907043457,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.011300318874418736,
+ "learning_rate": 0.002911302661146867,
+ "loss": 5.819517135620117,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.009915645234286785,
+ "learning_rate": 0.0029108483331254684,
+ "loss": 5.776939392089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.008965205401182175,
+ "learning_rate": 0.00291039288049428,
+ "loss": 5.824148178100586,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.011025927029550076,
+ "learning_rate": 0.0029099363036202515,
+ "loss": 5.832704544067383,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.010143411345779896,
+ "learning_rate": 0.0029094786028712387,
+ "loss": 5.802353382110596,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.010021256282925606,
+ "learning_rate": 0.002909019778616003,
+ "loss": 5.802136421203613,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.011356706731021404,
+ "learning_rate": 0.002908559831224211,
+ "loss": 5.807962417602539,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.010525298304855824,
+ "learning_rate": 0.002908098761066433,
+ "loss": 5.787661552429199,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.011147079057991505,
+ "learning_rate": 0.0029076365685141463,
+ "loss": 5.801015377044678,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.008777810260653496,
+ "learning_rate": 0.00290717325393973,
+ "loss": 5.779858589172363,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.011423609219491482,
+ "learning_rate": 0.002906708817716469,
+ "loss": 5.777166843414307,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.008639856241643429,
+ "learning_rate": 0.00290624326021855,
+ "loss": 5.747189521789551,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.009628898464143276,
+ "learning_rate": 0.002905776581821066,
+ "loss": 5.7666916847229,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.011988377198576927,
+ "learning_rate": 0.0029053087829000097,
+ "loss": 5.777951240539551,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.01693224348127842,
+ "learning_rate": 0.00290483986383228,
+ "loss": 5.818393707275391,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.020264796912670135,
+ "learning_rate": 0.002904369824995674,
+ "loss": 5.747611045837402,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.018283747136592865,
+ "learning_rate": 0.0029038986667688964,
+ "loss": 5.789663314819336,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.825676918029785,
+ "eval_runtime": 57.6908,
+ "eval_samples_per_second": 42.329,
+ "eval_steps_per_second": 1.335,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.011447477154433727,
+ "learning_rate": 0.002903426389531549,
+ "loss": 5.724975109100342,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.012539781630039215,
+ "learning_rate": 0.002902952993664138,
+ "loss": 5.747196674346924,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.012179837562143803,
+ "learning_rate": 0.002902478479548069,
+ "loss": 5.748995780944824,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.019896456971764565,
+ "learning_rate": 0.00290200284756565,
+ "loss": 5.751996994018555,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.022261742502450943,
+ "learning_rate": 0.002901526098100089,
+ "loss": 5.7516984939575195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02683536894619465,
+ "learning_rate": 0.0029010482315354946,
+ "loss": 5.716037750244141,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.018057072535157204,
+ "learning_rate": 0.0029005692482568747,
+ "loss": 5.778608322143555,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.011646891944110394,
+ "learning_rate": 0.002900089148650138,
+ "loss": 5.7308268547058105,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02230495773255825,
+ "learning_rate": 0.002899607933102092,
+ "loss": 5.768552780151367,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.01212464738637209,
+ "learning_rate": 0.002899125602000443,
+ "loss": 5.733015060424805,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.011764907278120518,
+ "learning_rate": 0.0028986421557337973,
+ "loss": 5.748104095458984,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.01593147963285446,
+ "learning_rate": 0.002898157594691658,
+ "loss": 5.701960563659668,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.016333773732185364,
+ "learning_rate": 0.0028976719192644277,
+ "loss": 5.716808795928955,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.017547868192195892,
+ "learning_rate": 0.002897185129843406,
+ "loss": 5.724377632141113,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.01662801019847393,
+ "learning_rate": 0.002896697226820791,
+ "loss": 5.737432956695557,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.015781691297888756,
+ "learning_rate": 0.002896208210589676,
+ "loss": 5.7376203536987305,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.014363214373588562,
+ "learning_rate": 0.0028957180815440534,
+ "loss": 5.7342610359191895,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.015048054046928883,
+ "learning_rate": 0.0028952268400788115,
+ "loss": 5.715073585510254,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.013753226026892662,
+ "learning_rate": 0.002894734486589734,
+ "loss": 5.73264217376709,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.015562563203275204,
+ "learning_rate": 0.002894241021473502,
+ "loss": 5.7072038650512695,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.01520819216966629,
+ "learning_rate": 0.0028937464451276916,
+ "loss": 5.683300018310547,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.013024724088609219,
+ "learning_rate": 0.002893250757950773,
+ "loss": 5.726273536682129,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.012760856188833714,
+ "learning_rate": 0.0028927539603421128,
+ "loss": 5.7249064445495605,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.013542579486966133,
+ "learning_rate": 0.0028922560527019723,
+ "loss": 5.70849084854126,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.014801754616200924,
+ "learning_rate": 0.002891757035431506,
+ "loss": 5.725185394287109,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.013000023551285267,
+ "learning_rate": 0.0028912569089327637,
+ "loss": 5.705799102783203,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.012919292785227299,
+ "learning_rate": 0.002890755673608688,
+ "loss": 5.674324989318848,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.015904201194643974,
+ "learning_rate": 0.002890253329863115,
+ "loss": 5.680788993835449,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.015467721968889236,
+ "learning_rate": 0.0028897498781007744,
+ "loss": 5.707531929016113,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.015157260000705719,
+ "learning_rate": 0.002889245318727288,
+ "loss": 5.689894676208496,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.014204306527972221,
+ "learning_rate": 0.00288873965214917,
+ "loss": 5.713001251220703,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.022045845165848732,
+ "learning_rate": 0.0028882328787738273,
+ "loss": 5.657631874084473,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.013380175456404686,
+ "learning_rate": 0.0028877249990095573,
+ "loss": 5.702192306518555,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.016698408871889114,
+ "learning_rate": 0.0028872160132655503,
+ "loss": 5.695206642150879,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.020001472905278206,
+ "learning_rate": 0.0028867059219518866,
+ "loss": 5.731742858886719,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.027350103482604027,
+ "learning_rate": 0.0028861947254795375,
+ "loss": 5.712314605712891,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.021862946450710297,
+ "learning_rate": 0.002885682424260365,
+ "loss": 5.697043418884277,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.019216977059841156,
+ "learning_rate": 0.0028851690187071204,
+ "loss": 5.698813438415527,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.02131095714867115,
+ "learning_rate": 0.002884654509233446,
+ "loss": 5.6918864250183105,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.016583887860178947,
+ "learning_rate": 0.0028841388962538724,
+ "loss": 5.711456298828125,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.016998684033751488,
+ "learning_rate": 0.0028836221801838194,
+ "loss": 5.675747871398926,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.01705755852162838,
+ "learning_rate": 0.0028831043614395966,
+ "loss": 5.697177410125732,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.014604352414608002,
+ "learning_rate": 0.0028825854404384,
+ "loss": 5.681950092315674,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.013561096973717213,
+ "learning_rate": 0.0028820654175983162,
+ "loss": 5.67357063293457,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.021149173378944397,
+ "learning_rate": 0.0028815442933383174,
+ "loss": 5.709477424621582,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.01431041955947876,
+ "learning_rate": 0.0028810220680782644,
+ "loss": 5.7012200355529785,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.01429709978401661,
+ "learning_rate": 0.0028804987422389045,
+ "loss": 5.689794540405273,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.032025713473558426,
+ "learning_rate": 0.0028799743162418712,
+ "loss": 5.6713762283325195,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.013168714940547943,
+ "learning_rate": 0.002879448790509686,
+ "loss": 5.685478210449219,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.014088666997849941,
+ "learning_rate": 0.0028789221654657556,
+ "loss": 5.645514965057373,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.013320053927600384,
+ "learning_rate": 0.002878394441534372,
+ "loss": 5.678678512573242,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.022180087864398956,
+ "learning_rate": 0.002877865619140712,
+ "loss": 5.660525321960449,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.01271889265626669,
+ "learning_rate": 0.002877335698710839,
+ "loss": 5.652164459228516,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.011617431417107582,
+ "learning_rate": 0.002876804680671701,
+ "loss": 5.6389946937561035,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.012040726840496063,
+ "learning_rate": 0.0028762725654511282,
+ "loss": 5.628744602203369,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.013148069381713867,
+ "learning_rate": 0.0028757393534778377,
+ "loss": 5.69610595703125,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.01826276071369648,
+ "learning_rate": 0.0028752050451814276,
+ "loss": 5.634238243103027,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.020393460988998413,
+ "learning_rate": 0.0028746696409923805,
+ "loss": 5.656968593597412,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.017072148621082306,
+ "learning_rate": 0.0028741331413420627,
+ "loss": 5.668219089508057,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.01694098301231861,
+ "learning_rate": 0.0028735955466627217,
+ "loss": 5.670073509216309,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.022760670632123947,
+ "learning_rate": 0.002873056857387488,
+ "loss": 5.649288177490234,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.03613442927598953,
+ "learning_rate": 0.0028725170739503734,
+ "loss": 5.723211288452148,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.04232329502701759,
+ "learning_rate": 0.002871976196786272,
+ "loss": 5.757920742034912,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.044210851192474365,
+ "learning_rate": 0.002871434226330959,
+ "loss": 5.702426910400391,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.044886402785778046,
+ "learning_rate": 0.0028708911630210905,
+ "loss": 5.742881774902344,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.022805195301771164,
+ "learning_rate": 0.002870347007294202,
+ "loss": 5.716353416442871,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02219395525753498,
+ "learning_rate": 0.00286980175958871,
+ "loss": 5.68503475189209,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.019240057095885277,
+ "learning_rate": 0.002869255420343911,
+ "loss": 5.700656890869141,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.01957295648753643,
+ "learning_rate": 0.0028687079899999807,
+ "loss": 5.70595645904541,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016326555982232094,
+ "learning_rate": 0.0028681594689979736,
+ "loss": 5.697229385375977,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015815885737538338,
+ "learning_rate": 0.0028676098577798234,
+ "loss": 5.7020792961120605,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.015865175053477287,
+ "learning_rate": 0.002867059156788342,
+ "loss": 5.648272514343262,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.7417449951171875,
+ "eval_runtime": 51.2263,
+ "eval_samples_per_second": 47.671,
+ "eval_steps_per_second": 1.503,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.016063183546066284,
+ "learning_rate": 0.0028665073664672185,
+ "loss": 5.65411376953125,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.014323881827294827,
+ "learning_rate": 0.0028659544872610214,
+ "loss": 5.629649639129639,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.014005238190293312,
+ "learning_rate": 0.002865400519615195,
+ "loss": 5.655135631561279,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.01385372132062912,
+ "learning_rate": 0.0028648454639760615,
+ "loss": 5.632213115692139,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.014073756523430347,
+ "learning_rate": 0.0028642893207908187,
+ "loss": 5.626431465148926,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.019551943987607956,
+ "learning_rate": 0.0028637320905075406,
+ "loss": 5.595447063446045,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.017571093514561653,
+ "learning_rate": 0.002863173773575179,
+ "loss": 5.661581039428711,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.01691303588449955,
+ "learning_rate": 0.0028626143704435585,
+ "loss": 5.641129493713379,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.01584242656826973,
+ "learning_rate": 0.0028620538815633807,
+ "loss": 5.640626907348633,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.014592235907912254,
+ "learning_rate": 0.0028614923073862213,
+ "loss": 5.611204147338867,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.016636405140161514,
+ "learning_rate": 0.0028609296483645304,
+ "loss": 5.614716053009033,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.01370833721011877,
+ "learning_rate": 0.002860365904951632,
+ "loss": 5.633011817932129,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.013218311592936516,
+ "learning_rate": 0.002859801077601725,
+ "loss": 5.588958740234375,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.016971616074442863,
+ "learning_rate": 0.0028592351667698786,
+ "loss": 5.635064125061035,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.015524905174970627,
+ "learning_rate": 0.0028586681729120393,
+ "loss": 5.611454963684082,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.014769157394766808,
+ "learning_rate": 0.0028581000964850215,
+ "loss": 5.62175178527832,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.014313343912363052,
+ "learning_rate": 0.002857530937946516,
+ "loss": 5.603628158569336,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.016748322173953056,
+ "learning_rate": 0.002856960697755083,
+ "loss": 5.617765426635742,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.013503300957381725,
+ "learning_rate": 0.0028563893763701538,
+ "loss": 5.636135578155518,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.012275655753910542,
+ "learning_rate": 0.002855816974252033,
+ "loss": 5.552533149719238,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.01288773212581873,
+ "learning_rate": 0.002855243491861894,
+ "loss": 5.620527267456055,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.013687008060514927,
+ "learning_rate": 0.002854668929661781,
+ "loss": 5.618595123291016,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.015678439289331436,
+ "learning_rate": 0.0028540932881146092,
+ "loss": 5.604031562805176,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.013567988760769367,
+ "learning_rate": 0.002853516567684162,
+ "loss": 5.549149513244629,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0157297495752573,
+ "learning_rate": 0.002852938768835092,
+ "loss": 5.618505477905273,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.011867019347846508,
+ "learning_rate": 0.0028523598920329233,
+ "loss": 5.579049110412598,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.01398563478142023,
+ "learning_rate": 0.0028517799377440443,
+ "loss": 5.597678184509277,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.015525273978710175,
+ "learning_rate": 0.0028511989064357148,
+ "loss": 5.601663589477539,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.01588466763496399,
+ "learning_rate": 0.002850616798576061,
+ "loss": 5.5650434494018555,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.014276969246566296,
+ "learning_rate": 0.0028500336146340772,
+ "loss": 5.607125282287598,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.013253689743578434,
+ "learning_rate": 0.002849449355079623,
+ "loss": 5.6299824714660645,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.01176661066710949,
+ "learning_rate": 0.0028488640203834263,
+ "loss": 5.560839653015137,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.011712565086781979,
+ "learning_rate": 0.0028482776110170813,
+ "loss": 5.618312835693359,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.013515647500753403,
+ "learning_rate": 0.002847690127453047,
+ "loss": 5.566764831542969,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.020177524536848068,
+ "learning_rate": 0.0028471015701646485,
+ "loss": 5.554571628570557,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.029197178781032562,
+ "learning_rate": 0.0028465119396260757,
+ "loss": 5.6097259521484375,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.027947140857577324,
+ "learning_rate": 0.0028459212363123824,
+ "loss": 5.65598201751709,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.021089809015393257,
+ "learning_rate": 0.0028453294606994887,
+ "loss": 5.67195987701416,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.019936271011829376,
+ "learning_rate": 0.0028447366132641773,
+ "loss": 5.609482765197754,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.0176444910466671,
+ "learning_rate": 0.002844142694484094,
+ "loss": 5.60432767868042,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.023154428228735924,
+ "learning_rate": 0.0028435477048377495,
+ "loss": 5.616270065307617,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.018391480669379234,
+ "learning_rate": 0.002842951644804515,
+ "loss": 5.600588321685791,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.013716557994484901,
+ "learning_rate": 0.002842354514864627,
+ "loss": 5.595095634460449,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.014372860081493855,
+ "learning_rate": 0.002841756315499181,
+ "loss": 5.584517478942871,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.014930845238268375,
+ "learning_rate": 0.002841157047190135,
+ "loss": 5.587005138397217,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.012631502002477646,
+ "learning_rate": 0.0028405567104203095,
+ "loss": 5.594094276428223,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.014677002094686031,
+ "learning_rate": 0.0028399553056733854,
+ "loss": 5.569582939147949,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.015558961778879166,
+ "learning_rate": 0.0028393528334339025,
+ "loss": 5.596535682678223,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.014322095550596714,
+ "learning_rate": 0.0028387492941872624,
+ "loss": 5.570324420928955,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.012913414277136326,
+ "learning_rate": 0.002838144688419726,
+ "loss": 5.573002338409424,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.01251310296356678,
+ "learning_rate": 0.0028375390166184137,
+ "loss": 5.550327777862549,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.013847017660737038,
+ "learning_rate": 0.0028369322792713026,
+ "loss": 5.54567813873291,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.013789358548820019,
+ "learning_rate": 0.0028363244768672313,
+ "loss": 5.571495056152344,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.01236631814390421,
+ "learning_rate": 0.0028357156098958954,
+ "loss": 5.569730758666992,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.013210169039666653,
+ "learning_rate": 0.002835105678847847,
+ "loss": 5.53984260559082,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.01604137383401394,
+ "learning_rate": 0.0028344946842144975,
+ "loss": 5.530598163604736,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.01708815060555935,
+ "learning_rate": 0.002833882626488114,
+ "loss": 5.529256820678711,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.017222166061401367,
+ "learning_rate": 0.0028332695061618197,
+ "loss": 5.564949035644531,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.01803218014538288,
+ "learning_rate": 0.002832655323729595,
+ "loss": 5.560049057006836,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.01658225618302822,
+ "learning_rate": 0.002832040079686276,
+ "loss": 5.553183555603027,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.016834143549203873,
+ "learning_rate": 0.002831423774527554,
+ "loss": 5.573678970336914,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.01856408454477787,
+ "learning_rate": 0.002830806408749974,
+ "loss": 5.539497375488281,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.026909375563263893,
+ "learning_rate": 0.002830187982850936,
+ "loss": 5.551433563232422,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.016046348959207535,
+ "learning_rate": 0.002829568497328696,
+ "loss": 5.566728591918945,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.018582617864012718,
+ "learning_rate": 0.0028289479526823614,
+ "loss": 5.543226718902588,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.0211537666618824,
+ "learning_rate": 0.0028283263494118943,
+ "loss": 5.560711860656738,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.013894101604819298,
+ "learning_rate": 0.0028277036880181095,
+ "loss": 5.528890132904053,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.017581885680556297,
+ "learning_rate": 0.0028270799690026725,
+ "loss": 5.57473087310791,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.01622696965932846,
+ "learning_rate": 0.002826455192868104,
+ "loss": 5.578961372375488,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.01966337114572525,
+ "learning_rate": 0.0028258293601177743,
+ "loss": 5.513942718505859,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021581366658210754,
+ "learning_rate": 0.002825202471255906,
+ "loss": 5.573669910430908,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.020468758419156075,
+ "learning_rate": 0.002824574526787571,
+ "loss": 5.516103744506836,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.626922130584717,
+ "eval_runtime": 54.2404,
+ "eval_samples_per_second": 45.022,
+ "eval_steps_per_second": 1.42,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.014597049914300442,
+ "learning_rate": 0.0028239455272186936,
+ "loss": 5.506453514099121,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.020277054980397224,
+ "learning_rate": 0.002823315473056048,
+ "loss": 5.535350799560547,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.019103925675153732,
+ "learning_rate": 0.0028226843648072562,
+ "loss": 5.522078037261963,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.07158710062503815,
+ "learning_rate": 0.0028220522029807913,
+ "loss": 5.531142234802246,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01958007924258709,
+ "learning_rate": 0.0028214189880859746,
+ "loss": 5.484055042266846,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.0185184795409441,
+ "learning_rate": 0.002820784720632976,
+ "loss": 5.511264801025391,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01976991444826126,
+ "learning_rate": 0.0028201494011328133,
+ "loss": 5.51821231842041,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.02188604325056076,
+ "learning_rate": 0.002819513030097352,
+ "loss": 5.483095169067383,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.019405223429203033,
+ "learning_rate": 0.002818875608039305,
+ "loss": 5.539701461791992,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.016907615587115288,
+ "learning_rate": 0.0028182371354722306,
+ "loss": 5.518251419067383,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.02160538360476494,
+ "learning_rate": 0.0028175976129105364,
+ "loss": 5.509786605834961,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.023125721141695976,
+ "learning_rate": 0.002816957040869473,
+ "loss": 5.535561561584473,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.022233380004763603,
+ "learning_rate": 0.0028163154198651372,
+ "loss": 5.553831100463867,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.017227552831172943,
+ "learning_rate": 0.002815672750414473,
+ "loss": 5.547080993652344,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.017122473567724228,
+ "learning_rate": 0.0028150290330352665,
+ "loss": 5.493778228759766,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.020641913637518883,
+ "learning_rate": 0.0028143842682461498,
+ "loss": 5.536645889282227,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.017383966594934464,
+ "learning_rate": 0.0028137384565665967,
+ "loss": 5.506233215332031,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.017969606444239616,
+ "learning_rate": 0.002813091598516928,
+ "loss": 5.524741172790527,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.01936613954603672,
+ "learning_rate": 0.0028124436946183047,
+ "loss": 5.460989952087402,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.017492495477199554,
+ "learning_rate": 0.0028117947453927313,
+ "loss": 5.487619876861572,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.018786396831274033,
+ "learning_rate": 0.0028111447513630545,
+ "loss": 5.516829013824463,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.020149117335677147,
+ "learning_rate": 0.0028104937130529624,
+ "loss": 5.460541725158691,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.024668103083968163,
+ "learning_rate": 0.0028098416309869857,
+ "loss": 5.497687339782715,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.015888608992099762,
+ "learning_rate": 0.002809188505690494,
+ "loss": 5.494459629058838,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.019327623769640923,
+ "learning_rate": 0.0028085343376896997,
+ "loss": 5.486351490020752,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.026855193078517914,
+ "learning_rate": 0.0028078791275116532,
+ "loss": 5.509825706481934,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.026340370997786522,
+ "learning_rate": 0.002807222875684246,
+ "loss": 5.527161598205566,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.02200835384428501,
+ "learning_rate": 0.002806565582736208,
+ "loss": 5.507146835327148,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.020420534536242485,
+ "learning_rate": 0.0028059072491971085,
+ "loss": 5.5162506103515625,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02493203990161419,
+ "learning_rate": 0.0028052478755973545,
+ "loss": 5.5252556800842285,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.021293973550200462,
+ "learning_rate": 0.0028045874624681925,
+ "loss": 5.523589134216309,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.016593236476182938,
+ "learning_rate": 0.0028039260103417045,
+ "loss": 5.500586032867432,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.018857723101973534,
+ "learning_rate": 0.002803263519750811,
+ "loss": 5.52933406829834,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.015930436551570892,
+ "learning_rate": 0.0028025999912292677,
+ "loss": 5.48813533782959,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.019427936524152756,
+ "learning_rate": 0.002801935425311669,
+ "loss": 5.46573543548584,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.018123053014278412,
+ "learning_rate": 0.0028012698225334427,
+ "loss": 5.499445915222168,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.02035410702228546,
+ "learning_rate": 0.002800603183430853,
+ "loss": 5.513540267944336,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.025838350877165794,
+ "learning_rate": 0.0027999355085409996,
+ "loss": 5.565637588500977,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.029621612280607224,
+ "learning_rate": 0.002799266798401816,
+ "loss": 5.5619215965271,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.028948379680514336,
+ "learning_rate": 0.0027985970535520685,
+ "loss": 5.551331520080566,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.022135930135846138,
+ "learning_rate": 0.0027979262745313604,
+ "loss": 5.527722358703613,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.028884069994091988,
+ "learning_rate": 0.0027972544618801256,
+ "loss": 5.554599761962891,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.02499174140393734,
+ "learning_rate": 0.00279658161613963,
+ "loss": 5.543351173400879,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.021131042391061783,
+ "learning_rate": 0.0027959077378519755,
+ "loss": 5.496039390563965,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.02957489714026451,
+ "learning_rate": 0.002795232827560092,
+ "loss": 5.552300453186035,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.022832119837403297,
+ "learning_rate": 0.0027945568858077433,
+ "loss": 5.488651275634766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.02074701525270939,
+ "learning_rate": 0.002793879913139523,
+ "loss": 5.491244792938232,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.019919803366065025,
+ "learning_rate": 0.002793201910100856,
+ "loss": 5.483855247497559,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.01858169585466385,
+ "learning_rate": 0.002792522877237998,
+ "loss": 5.50787353515625,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02085857465863228,
+ "learning_rate": 0.002791842815098032,
+ "loss": 5.512872695922852,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.01407450158149004,
+ "learning_rate": 0.002791161724228872,
+ "loss": 5.472590446472168,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.01672944612801075,
+ "learning_rate": 0.002790479605179262,
+ "loss": 5.469183921813965,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.016565700992941856,
+ "learning_rate": 0.0027897964584987713,
+ "loss": 5.493677139282227,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.017017360776662827,
+ "learning_rate": 0.0027891122847377996,
+ "loss": 5.464931488037109,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.019842486828565598,
+ "learning_rate": 0.002788427084447574,
+ "loss": 5.461293697357178,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.016210904344916344,
+ "learning_rate": 0.0027877408581801467,
+ "loss": 5.454716682434082,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.02064039371907711,
+ "learning_rate": 0.002787053606488398,
+ "loss": 5.466576099395752,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.018198711797595024,
+ "learning_rate": 0.0027863653299260356,
+ "loss": 5.447747230529785,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.015937626361846924,
+ "learning_rate": 0.00278567602904759,
+ "loss": 5.5055036544799805,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.01687259040772915,
+ "learning_rate": 0.002784985704408419,
+ "loss": 5.422250747680664,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.02622046135365963,
+ "learning_rate": 0.002784294356564705,
+ "loss": 5.466894626617432,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.018672186881303787,
+ "learning_rate": 0.0027836019860734535,
+ "loss": 5.48300838470459,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.02166241593658924,
+ "learning_rate": 0.0027829085934924958,
+ "loss": 5.477658271789551,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018426023423671722,
+ "learning_rate": 0.0027822141793804856,
+ "loss": 5.487186431884766,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.022229891270399094,
+ "learning_rate": 0.0027815187442968995,
+ "loss": 5.488579273223877,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.026930972933769226,
+ "learning_rate": 0.002780822288802037,
+ "loss": 5.492000102996826,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.02733265422284603,
+ "learning_rate": 0.00278012481345702,
+ "loss": 5.4663801193237305,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.02993253618478775,
+ "learning_rate": 0.0027794263188237922,
+ "loss": 5.543368816375732,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03152153640985489,
+ "learning_rate": 0.002778726805465118,
+ "loss": 5.565644264221191,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.031395044177770615,
+ "learning_rate": 0.002778026273944581,
+ "loss": 5.553359508514404,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.026603225618600845,
+ "learning_rate": 0.002777324724826589,
+ "loss": 5.552347183227539,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.024152139201760292,
+ "learning_rate": 0.002776622158676366,
+ "loss": 5.493257522583008,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.602636814117432,
+ "eval_runtime": 54.6199,
+ "eval_samples_per_second": 44.709,
+ "eval_steps_per_second": 1.41,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.02089761756360531,
+ "learning_rate": 0.0027759185760599576,
+ "loss": 5.461747646331787,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.018181778490543365,
+ "learning_rate": 0.0027752139775442278,
+ "loss": 5.492741584777832,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.020166238769888878,
+ "learning_rate": 0.0027745083636968575,
+ "loss": 5.476003646850586,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.015491694211959839,
+ "learning_rate": 0.0027738017350863488,
+ "loss": 5.425764560699463,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.022291697561740875,
+ "learning_rate": 0.0027730940922820182,
+ "loss": 5.45866584777832,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.018074683845043182,
+ "learning_rate": 0.002772385435854002,
+ "loss": 5.465143203735352,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.016872268170118332,
+ "learning_rate": 0.002771675766373251,
+ "loss": 5.446003437042236,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.023103339597582817,
+ "learning_rate": 0.0027709650844115336,
+ "loss": 5.411036491394043,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.027957158163189888,
+ "learning_rate": 0.0027702533905414324,
+ "loss": 5.452442169189453,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.019935399293899536,
+ "learning_rate": 0.002769540685336348,
+ "loss": 5.412145614624023,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.014351371675729752,
+ "learning_rate": 0.0027688269693704926,
+ "loss": 5.434524059295654,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.0152254244312644,
+ "learning_rate": 0.0027681122432188948,
+ "loss": 5.428544998168945,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.015071330592036247,
+ "learning_rate": 0.0027673965074573965,
+ "loss": 5.459929466247559,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.023415958508849144,
+ "learning_rate": 0.0027666797626626537,
+ "loss": 5.447159767150879,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.017228802666068077,
+ "learning_rate": 0.0027659620094121337,
+ "loss": 5.442273139953613,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.014184504747390747,
+ "learning_rate": 0.002765243248284118,
+ "loss": 5.448305130004883,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.024384526535868645,
+ "learning_rate": 0.0027645234798576988,
+ "loss": 5.4076337814331055,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.043152231723070145,
+ "learning_rate": 0.0027638027047127813,
+ "loss": 5.45344352722168,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.02889758162200451,
+ "learning_rate": 0.002763080923430081,
+ "loss": 5.489194393157959,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.02686968818306923,
+ "learning_rate": 0.002762358136591123,
+ "loss": 5.493659496307373,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.020467868074774742,
+ "learning_rate": 0.002761634344778245,
+ "loss": 5.410489559173584,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.019759224727749825,
+ "learning_rate": 0.002760909548574592,
+ "loss": 5.439385890960693,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.0233843382447958,
+ "learning_rate": 0.0027601837485641194,
+ "loss": 5.445036888122559,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.031753458082675934,
+ "learning_rate": 0.0027594569453315916,
+ "loss": 5.503895282745361,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.02089349366724491,
+ "learning_rate": 0.002758729139462581,
+ "loss": 5.45689582824707,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.03339357674121857,
+ "learning_rate": 0.002758000331543466,
+ "loss": 5.522361755371094,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019328996539115906,
+ "learning_rate": 0.002757270522161436,
+ "loss": 5.441373825073242,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019863856956362724,
+ "learning_rate": 0.002756539711904484,
+ "loss": 5.4633283615112305,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.02141754701733589,
+ "learning_rate": 0.0027558079013614116,
+ "loss": 5.453215599060059,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016591915860772133,
+ "learning_rate": 0.002755075091121824,
+ "loss": 5.463696479797363,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.01955028623342514,
+ "learning_rate": 0.0027543412817761345,
+ "loss": 5.411355018615723,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.016791662201285362,
+ "learning_rate": 0.0027536064739155596,
+ "loss": 5.459614276885986,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.019661981612443924,
+ "learning_rate": 0.00275287066813212,
+ "loss": 5.440699577331543,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.01927965320646763,
+ "learning_rate": 0.002752133865018642,
+ "loss": 5.409865856170654,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.016483580693602562,
+ "learning_rate": 0.002751396065168754,
+ "loss": 5.390999794006348,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016264773905277252,
+ "learning_rate": 0.0027506572691768888,
+ "loss": 5.418543815612793,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.01958460919559002,
+ "learning_rate": 0.00274991747763828,
+ "loss": 5.403289794921875,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.01861695572733879,
+ "learning_rate": 0.002749176691148965,
+ "loss": 5.431911468505859,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.029594263061881065,
+ "learning_rate": 0.002748434910305782,
+ "loss": 5.399242401123047,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.028574613854289055,
+ "learning_rate": 0.0027476921357063697,
+ "loss": 5.440446853637695,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.028612585738301277,
+ "learning_rate": 0.0027469483679491693,
+ "loss": 5.410346031188965,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.020704898983240128,
+ "learning_rate": 0.00274620360763342,
+ "loss": 5.44694709777832,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.020542364567518234,
+ "learning_rate": 0.002745457855359163,
+ "loss": 5.402653694152832,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.015762662515044212,
+ "learning_rate": 0.0027447111117272363,
+ "loss": 5.396594047546387,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022096984088420868,
+ "learning_rate": 0.002743963377339278,
+ "loss": 5.398744106292725,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.018791764974594116,
+ "learning_rate": 0.002743214652797724,
+ "loss": 5.397758483886719,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01642526686191559,
+ "learning_rate": 0.0027424649387058093,
+ "loss": 5.411459922790527,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.017931703478097916,
+ "learning_rate": 0.0027417142356675637,
+ "loss": 5.390264511108398,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018819602206349373,
+ "learning_rate": 0.0027409625442878153,
+ "loss": 5.407539367675781,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.019389411434531212,
+ "learning_rate": 0.0027402098651721886,
+ "loss": 5.386474609375,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.013645983301103115,
+ "learning_rate": 0.002739456198927104,
+ "loss": 5.373845100402832,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.013846895657479763,
+ "learning_rate": 0.002738701546159776,
+ "loss": 5.387155055999756,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013553125783801079,
+ "learning_rate": 0.002737945907478215,
+ "loss": 5.377801895141602,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.012771612964570522,
+ "learning_rate": 0.002737189283491225,
+ "loss": 5.344066143035889,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.020465301349759102,
+ "learning_rate": 0.002736431674808405,
+ "loss": 5.382771968841553,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.014798106625676155,
+ "learning_rate": 0.0027356730820401465,
+ "loss": 5.370522499084473,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.01566634140908718,
+ "learning_rate": 0.0027349135057976332,
+ "loss": 5.351980209350586,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.017485102638602257,
+ "learning_rate": 0.002734152946692843,
+ "loss": 5.355443477630615,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.016646096482872963,
+ "learning_rate": 0.0027333914053385444,
+ "loss": 5.34829044342041,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.017293352633714676,
+ "learning_rate": 0.0027326288823482964,
+ "loss": 5.3519487380981445,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.018873002380132675,
+ "learning_rate": 0.002731865378336452,
+ "loss": 5.348442077636719,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.015504784882068634,
+ "learning_rate": 0.002731100893918151,
+ "loss": 5.341490268707275,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.014270870015025139,
+ "learning_rate": 0.0027303354297093254,
+ "loss": 5.359251022338867,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.018820680677890778,
+ "learning_rate": 0.002729568986326695,
+ "loss": 5.390280723571777,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.021008187904953957,
+ "learning_rate": 0.0027288015643877708,
+ "loss": 5.338380336761475,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.016266655176877975,
+ "learning_rate": 0.0027280331645108498,
+ "loss": 5.347043037414551,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.01831045374274254,
+ "learning_rate": 0.0027272637873150185,
+ "loss": 5.326537132263184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02073794975876808,
+ "learning_rate": 0.002726493433420149,
+ "loss": 5.3607025146484375,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.022534754127264023,
+ "learning_rate": 0.0027257221034469033,
+ "loss": 5.353292465209961,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.021627362817525864,
+ "learning_rate": 0.002724949798016727,
+ "loss": 5.373323917388916,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.018226243555545807,
+ "learning_rate": 0.0027241765177518523,
+ "loss": 5.344578266143799,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.02032875455915928,
+ "learning_rate": 0.002723402263275298,
+ "loss": 5.329159259796143,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.44969367980957,
+ "eval_runtime": 51.2294,
+ "eval_samples_per_second": 47.668,
+ "eval_steps_per_second": 1.503,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.01613701693713665,
+ "learning_rate": 0.0027226270352108664,
+ "loss": 5.3024678230285645,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01940973661839962,
+ "learning_rate": 0.0027218508341831445,
+ "loss": 5.270181179046631,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.015840493142604828,
+ "learning_rate": 0.002721073660817504,
+ "loss": 5.261706352233887,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.017390675842761993,
+ "learning_rate": 0.0027202955157401,
+ "loss": 5.262207984924316,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.015184012241661549,
+ "learning_rate": 0.0027195163995778686,
+ "loss": 5.297155380249023,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.013417073525488377,
+ "learning_rate": 0.00271873631295853,
+ "loss": 5.279838562011719,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.013024317100644112,
+ "learning_rate": 0.0027179552565105864,
+ "loss": 5.265446186065674,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.01426338218152523,
+ "learning_rate": 0.002717173230863321,
+ "loss": 5.284180164337158,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.019604040309786797,
+ "learning_rate": 0.002716390236646797,
+ "loss": 5.286281108856201,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.019039476290345192,
+ "learning_rate": 0.0027156062744918587,
+ "loss": 5.290433406829834,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.019284334033727646,
+ "learning_rate": 0.0027148213450301313,
+ "loss": 5.262433052062988,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.0145154669880867,
+ "learning_rate": 0.0027140354488940175,
+ "loss": 5.24387264251709,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.014776417054235935,
+ "learning_rate": 0.0027132485867166992,
+ "loss": 5.282435417175293,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.012711729854345322,
+ "learning_rate": 0.0027124607591321377,
+ "loss": 5.246735572814941,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.013394835405051708,
+ "learning_rate": 0.0027116719667750716,
+ "loss": 5.270739555358887,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.013468850404024124,
+ "learning_rate": 0.002710882210281016,
+ "loss": 5.27161169052124,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.030262840911746025,
+ "learning_rate": 0.002710091490286264,
+ "loss": 5.256673812866211,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.013560288585722446,
+ "learning_rate": 0.0027092998074278846,
+ "loss": 5.253674507141113,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014592813327908516,
+ "learning_rate": 0.002708507162343721,
+ "loss": 5.25804328918457,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014328169636428356,
+ "learning_rate": 0.0027077135556723953,
+ "loss": 5.235117435455322,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.01277366653084755,
+ "learning_rate": 0.0027069189880533,
+ "loss": 5.2053985595703125,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.012948785908520222,
+ "learning_rate": 0.0027061234601266043,
+ "loss": 5.214653015136719,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.014892828650772572,
+ "learning_rate": 0.002705326972533252,
+ "loss": 5.227547645568848,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01427230704575777,
+ "learning_rate": 0.002704529525914957,
+ "loss": 5.241065502166748,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.01650116965174675,
+ "learning_rate": 0.0027037311209142092,
+ "loss": 5.227713108062744,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.019728578627109528,
+ "learning_rate": 0.0027029317581742677,
+ "loss": 5.2624921798706055,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.023847149685025215,
+ "learning_rate": 0.0027021314383391657,
+ "loss": 5.278811454772949,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.020690469071269035,
+ "learning_rate": 0.0027013301620537056,
+ "loss": 5.238919258117676,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.021211959421634674,
+ "learning_rate": 0.0027005279299634616,
+ "loss": 5.253684997558594,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.025854922831058502,
+ "learning_rate": 0.002699724742714778,
+ "loss": 5.2538652420043945,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.0215485580265522,
+ "learning_rate": 0.0026989206009547678,
+ "loss": 5.224540710449219,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.019827792420983315,
+ "learning_rate": 0.0026981155053313137,
+ "loss": 5.220674514770508,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.017689093947410583,
+ "learning_rate": 0.002697309456493066,
+ "loss": 5.228511810302734,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.022564806044101715,
+ "learning_rate": 0.002696502455089445,
+ "loss": 5.295933723449707,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.020070280879735947,
+ "learning_rate": 0.0026956945017706357,
+ "loss": 5.230920791625977,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.025744661688804626,
+ "learning_rate": 0.002694885597187593,
+ "loss": 5.296453952789307,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.021263007074594498,
+ "learning_rate": 0.0026940757419920357,
+ "loss": 5.220481872558594,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.017577245831489563,
+ "learning_rate": 0.0026932649368364495,
+ "loss": 5.231488227844238,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.015562832355499268,
+ "learning_rate": 0.002692453182374086,
+ "loss": 5.175224304199219,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.014115403406322002,
+ "learning_rate": 0.0026916404792589606,
+ "loss": 5.228387832641602,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.012863830663263798,
+ "learning_rate": 0.002690826828145854,
+ "loss": 5.186727523803711,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.014151126146316528,
+ "learning_rate": 0.002690012229690309,
+ "loss": 5.1694817543029785,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.01316932775080204,
+ "learning_rate": 0.0026891966845486353,
+ "loss": 5.19931697845459,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.011340726166963577,
+ "learning_rate": 0.0026883801933779003,
+ "loss": 5.238619804382324,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.012616089545190334,
+ "learning_rate": 0.0026875627568359378,
+ "loss": 5.183376789093018,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.012274038046598434,
+ "learning_rate": 0.002686744375581341,
+ "loss": 5.199946403503418,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.014990398660302162,
+ "learning_rate": 0.002685925050273465,
+ "loss": 5.168451309204102,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.011399757117033005,
+ "learning_rate": 0.0026851047815724255,
+ "loss": 5.173828125,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.013638390228152275,
+ "learning_rate": 0.0026842835701390982,
+ "loss": 5.142702102661133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.014472718350589275,
+ "learning_rate": 0.0026834614166351177,
+ "loss": 5.173436164855957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.014509808272123337,
+ "learning_rate": 0.002682638321722879,
+ "loss": 5.173975467681885,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.013576583005487919,
+ "learning_rate": 0.0026818142860655352,
+ "loss": 5.1435089111328125,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.012914472259581089,
+ "learning_rate": 0.0026809893103269954,
+ "loss": 5.151381492614746,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.013803352601826191,
+ "learning_rate": 0.0026801633951719288,
+ "loss": 5.133159637451172,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.018488213419914246,
+ "learning_rate": 0.0026793365412657606,
+ "loss": 5.183903694152832,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.01843969337642193,
+ "learning_rate": 0.002678508749274671,
+ "loss": 5.1660356521606445,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.017181839793920517,
+ "learning_rate": 0.0026776800198655983,
+ "loss": 5.179847717285156,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0148160420358181,
+ "learning_rate": 0.0026768503537062333,
+ "loss": 5.149198532104492,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.013251296244561672,
+ "learning_rate": 0.0026760197514650247,
+ "loss": 5.192000389099121,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.010746927931904793,
+ "learning_rate": 0.0026751882138111717,
+ "loss": 5.138680458068848,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.012316994369029999,
+ "learning_rate": 0.002674355741414631,
+ "loss": 5.169965744018555,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.01191797386854887,
+ "learning_rate": 0.0026735223349461106,
+ "loss": 5.128267288208008,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.013525084592401981,
+ "learning_rate": 0.0026726879950770698,
+ "loss": 5.124795436859131,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.014843370765447617,
+ "learning_rate": 0.002671852722479721,
+ "loss": 5.1497087478637695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.015285412780940533,
+ "learning_rate": 0.00267101651782703,
+ "loss": 5.093118190765381,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.013555832207202911,
+ "learning_rate": 0.0026701793817927107,
+ "loss": 5.13249397277832,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.010569742880761623,
+ "learning_rate": 0.0026693413150512273,
+ "loss": 5.084077835083008,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.011483813636004925,
+ "learning_rate": 0.0026685023182777975,
+ "loss": 5.128049850463867,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.013724366202950478,
+ "learning_rate": 0.0026676623921483844,
+ "loss": 5.157539367675781,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.014337432570755482,
+ "learning_rate": 0.0026668215373397016,
+ "loss": 5.149948596954346,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.014101454056799412,
+ "learning_rate": 0.00266597975452921,
+ "loss": 5.117250442504883,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.016004368662834167,
+ "learning_rate": 0.0026651370443951197,
+ "loss": 5.143221855163574,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.249833583831787,
+ "eval_runtime": 51.5543,
+ "eval_samples_per_second": 47.368,
+ "eval_steps_per_second": 1.494,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.01297403872013092,
+ "learning_rate": 0.0026642934076163867,
+ "loss": 5.089287281036377,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.012867718935012817,
+ "learning_rate": 0.0026634488448727144,
+ "loss": 5.0784101486206055,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.014600927010178566,
+ "learning_rate": 0.0026626033568445504,
+ "loss": 5.050220489501953,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.016619572415947914,
+ "learning_rate": 0.0026617569442130905,
+ "loss": 5.079303741455078,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.015560587868094444,
+ "learning_rate": 0.0026609096076602736,
+ "loss": 5.072385311126709,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.01774057187139988,
+ "learning_rate": 0.0026600613478687833,
+ "loss": 5.122529983520508,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.019971400499343872,
+ "learning_rate": 0.0026592121655220472,
+ "loss": 5.085966110229492,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.017667256295681,
+ "learning_rate": 0.0026583620613042365,
+ "loss": 5.078432559967041,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.022369660437107086,
+ "learning_rate": 0.002657511035900264,
+ "loss": 5.0630903244018555,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.017964458093047142,
+ "learning_rate": 0.002656659089995787,
+ "loss": 5.0846076011657715,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.021769145503640175,
+ "learning_rate": 0.002655806224277201,
+ "loss": 5.081344127655029,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.02487620897591114,
+ "learning_rate": 0.0026549524394316464,
+ "loss": 5.106287956237793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.022516218945384026,
+ "learning_rate": 0.0026540977361470012,
+ "loss": 5.128098487854004,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.02099926583468914,
+ "learning_rate": 0.0026532421151118845,
+ "loss": 5.103562355041504,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.01899387687444687,
+ "learning_rate": 0.002652385577015655,
+ "loss": 5.062675476074219,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.016801996156573296,
+ "learning_rate": 0.002651528122548409,
+ "loss": 5.096002578735352,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.014009200967848301,
+ "learning_rate": 0.0026506697524009836,
+ "loss": 5.075201034545898,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.014653858728706837,
+ "learning_rate": 0.002649810467264951,
+ "loss": 5.090478897094727,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.014918150380253792,
+ "learning_rate": 0.0026489502678326226,
+ "loss": 5.070797920227051,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.01327288057655096,
+ "learning_rate": 0.0026480891547970443,
+ "loss": 5.030500411987305,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.012885140255093575,
+ "learning_rate": 0.002647227128852,
+ "loss": 5.0845537185668945,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.012848512269556522,
+ "learning_rate": 0.002646364190692009,
+ "loss": 5.079484939575195,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.01149789709597826,
+ "learning_rate": 0.002645500341012324,
+ "loss": 5.036316394805908,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.010360757820308208,
+ "learning_rate": 0.0026446355805089337,
+ "loss": 5.04129695892334,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.010426290333271027,
+ "learning_rate": 0.00264376990987856,
+ "loss": 5.051270484924316,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.01018153689801693,
+ "learning_rate": 0.002642903329818658,
+ "loss": 5.102342128753662,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.011879283003509045,
+ "learning_rate": 0.0026420358410274153,
+ "loss": 5.053501129150391,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.010418329387903214,
+ "learning_rate": 0.002641167444203752,
+ "loss": 5.030653476715088,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.012302013114094734,
+ "learning_rate": 0.0026402981400473202,
+ "loss": 5.0402069091796875,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.011392916552722454,
+ "learning_rate": 0.002639427929258502,
+ "loss": 5.024111747741699,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.013075058348476887,
+ "learning_rate": 0.0026385568125384114,
+ "loss": 5.021731376647949,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015349832363426685,
+ "learning_rate": 0.0026376847905888903,
+ "loss": 5.03192138671875,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01588309183716774,
+ "learning_rate": 0.0026368118641125113,
+ "loss": 4.966363906860352,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.014378590509295464,
+ "learning_rate": 0.0026359380338125764,
+ "loss": 5.070773124694824,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.012306655757129192,
+ "learning_rate": 0.002635063300393114,
+ "loss": 5.025704860687256,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01212004292756319,
+ "learning_rate": 0.002634187664558881,
+ "loss": 4.965104103088379,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013538849540054798,
+ "learning_rate": 0.0026333111270153624,
+ "loss": 5.029810905456543,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.012052906677126884,
+ "learning_rate": 0.0026324336884687672,
+ "loss": 5.0262908935546875,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.011368660256266594,
+ "learning_rate": 0.002631555349626033,
+ "loss": 5.001485824584961,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.010215395130217075,
+ "learning_rate": 0.0026306761111948205,
+ "loss": 5.019739151000977,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.011449878104031086,
+ "learning_rate": 0.0026297959738835167,
+ "loss": 4.9816083908081055,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.012381811626255512,
+ "learning_rate": 0.0026289149384012326,
+ "loss": 5.025032043457031,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.00974641740322113,
+ "learning_rate": 0.002628033005457802,
+ "loss": 5.000177383422852,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.009515874087810516,
+ "learning_rate": 0.0026271501757637833,
+ "loss": 4.996979713439941,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.011125455610454082,
+ "learning_rate": 0.0026262664500304555,
+ "loss": 4.996796607971191,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.013290468603372574,
+ "learning_rate": 0.0026253818289698207,
+ "loss": 5.053703308105469,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.014650283381342888,
+ "learning_rate": 0.002624496313294602,
+ "loss": 5.001321792602539,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014537827111780643,
+ "learning_rate": 0.002623609903718244,
+ "loss": 5.024584770202637,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.012908566743135452,
+ "learning_rate": 0.0026227226009549105,
+ "loss": 5.042878150939941,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.01446808222681284,
+ "learning_rate": 0.002621834405719485,
+ "loss": 5.020857810974121,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.01535153016448021,
+ "learning_rate": 0.0026209453187275708,
+ "loss": 5.0056304931640625,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01815616339445114,
+ "learning_rate": 0.0026200553406954892,
+ "loss": 5.0467376708984375,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.023135516792535782,
+ "learning_rate": 0.0026191644723402795,
+ "loss": 5.033995628356934,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.019682930782437325,
+ "learning_rate": 0.002618272714379698,
+ "loss": 5.008149147033691,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.018521904945373535,
+ "learning_rate": 0.002617380067532217,
+ "loss": 5.007076263427734,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.02118663489818573,
+ "learning_rate": 0.0026164865325170278,
+ "loss": 5.05519962310791,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024684788659214973,
+ "learning_rate": 0.0026155921100540344,
+ "loss": 5.027571201324463,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.02609819918870926,
+ "learning_rate": 0.0026146968008638567,
+ "loss": 5.10361385345459,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.025958053767681122,
+ "learning_rate": 0.00261380060566783,
+ "loss": 5.0911455154418945,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.017067518085241318,
+ "learning_rate": 0.002612903525188002,
+ "loss": 5.045783042907715,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.016510486602783203,
+ "learning_rate": 0.0026120055601471345,
+ "loss": 5.069464683532715,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015552333556115627,
+ "learning_rate": 0.0026111067112687012,
+ "loss": 5.024248123168945,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.013432685285806656,
+ "learning_rate": 0.0026102069792768886,
+ "loss": 5.004837989807129,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.012300201691687107,
+ "learning_rate": 0.0026093063648965943,
+ "loss": 5.031093597412109,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.01197744719684124,
+ "learning_rate": 0.002608404868853428,
+ "loss": 4.965237617492676,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.01100246049463749,
+ "learning_rate": 0.002607502491873708,
+ "loss": 5.04719352722168,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.010111447423696518,
+ "learning_rate": 0.0026065992346844626,
+ "loss": 4.986410617828369,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.010669790208339691,
+ "learning_rate": 0.002605695098013431,
+ "loss": 4.9825663566589355,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.010805477388203144,
+ "learning_rate": 0.0026047900825890593,
+ "loss": 4.991506576538086,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.010886074975132942,
+ "learning_rate": 0.002603884189140502,
+ "loss": 4.987531661987305,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.010735839605331421,
+ "learning_rate": 0.0026029774183976206,
+ "loss": 4.982756614685059,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.011092593893408775,
+ "learning_rate": 0.002602069771090985,
+ "loss": 4.994824409484863,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 5.13364315032959,
+ "eval_runtime": 50.4195,
+ "eval_samples_per_second": 48.434,
+ "eval_steps_per_second": 1.527,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.01005905494093895,
+ "learning_rate": 0.0026011612479518694,
+ "loss": 4.918412685394287,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.009904682636260986,
+ "learning_rate": 0.002600251849712255,
+ "loss": 4.924630165100098,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.009543103165924549,
+ "learning_rate": 0.0025993415771048273,
+ "loss": 4.940309524536133,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.011031470261514187,
+ "learning_rate": 0.0025984304308629765,
+ "loss": 4.924609184265137,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.009511022828519344,
+ "learning_rate": 0.002597518411720796,
+ "loss": 4.9102325439453125,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.00886621605604887,
+ "learning_rate": 0.0025966055204130847,
+ "loss": 4.9236907958984375,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.010057039558887482,
+ "learning_rate": 0.002595691757675341,
+ "loss": 4.908633232116699,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.01098555512726307,
+ "learning_rate": 0.0025947771242437684,
+ "loss": 4.935927391052246,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.008800056762993336,
+ "learning_rate": 0.0025938616208552694,
+ "loss": 4.873291969299316,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.009747766889631748,
+ "learning_rate": 0.0025929452482474497,
+ "loss": 4.954436302185059,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.010785684920847416,
+ "learning_rate": 0.002592028007158613,
+ "loss": 4.924927711486816,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.012620717287063599,
+ "learning_rate": 0.0025911098983277645,
+ "loss": 4.953606605529785,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.014428738504648209,
+ "learning_rate": 0.002590190922494607,
+ "loss": 4.951706886291504,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.014496048912405968,
+ "learning_rate": 0.0025892710803995444,
+ "loss": 4.941249847412109,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.0109238987788558,
+ "learning_rate": 0.0025883503727836764,
+ "loss": 4.939975261688232,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.012271072715520859,
+ "learning_rate": 0.0025874288003887996,
+ "loss": 4.880942344665527,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.016302887350320816,
+ "learning_rate": 0.0025865063639574086,
+ "loss": 4.931602478027344,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.016137251630425453,
+ "learning_rate": 0.0025855830642326945,
+ "loss": 4.918013095855713,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.014732961542904377,
+ "learning_rate": 0.002584658901958542,
+ "loss": 4.927577495574951,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.016387946903705597,
+ "learning_rate": 0.0025837338778795325,
+ "loss": 4.930819034576416,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.017146151512861252,
+ "learning_rate": 0.002582807992740941,
+ "loss": 4.928377151489258,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.016153613105416298,
+ "learning_rate": 0.0025818812472887375,
+ "loss": 4.975607395172119,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.013386047445237637,
+ "learning_rate": 0.0025809536422695825,
+ "loss": 4.9434709548950195,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.012328253127634525,
+ "learning_rate": 0.0025800251784308314,
+ "loss": 4.917585372924805,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.012000586837530136,
+ "learning_rate": 0.0025790958565205316,
+ "loss": 4.9137725830078125,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.009463997557759285,
+ "learning_rate": 0.002578165677287419,
+ "loss": 4.910362243652344,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.010591360740363598,
+ "learning_rate": 0.0025772346414809242,
+ "loss": 4.916032314300537,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.010429762303829193,
+ "learning_rate": 0.002576302749851164,
+ "loss": 4.934561729431152,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.12526583671569824,
+ "learning_rate": 0.0025753700031489486,
+ "loss": 4.897515773773193,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01368892751634121,
+ "learning_rate": 0.0025744364021257743,
+ "loss": 4.94814920425415,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.016603967174887657,
+ "learning_rate": 0.0025735019475338254,
+ "loss": 4.8996758460998535,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.06110106781125069,
+ "learning_rate": 0.0025725666401259775,
+ "loss": 4.946794509887695,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.018292423337697983,
+ "learning_rate": 0.002571630480655789,
+ "loss": 4.958992958068848,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.01852245442569256,
+ "learning_rate": 0.0025706934698775073,
+ "loss": 4.942556381225586,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.0195649154484272,
+ "learning_rate": 0.0025697556085460646,
+ "loss": 4.895773887634277,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.017820244655013084,
+ "learning_rate": 0.00256881689741708,
+ "loss": 4.915250778198242,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.016814742237329483,
+ "learning_rate": 0.002567877337246855,
+ "loss": 4.9521260261535645,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.015319781377911568,
+ "learning_rate": 0.002566936928792377,
+ "loss": 4.899402141571045,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.01607401855289936,
+ "learning_rate": 0.0025659956728113146,
+ "loss": 4.909913063049316,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.015564955770969391,
+ "learning_rate": 0.002565053570062023,
+ "loss": 4.947719573974609,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.017502037808299065,
+ "learning_rate": 0.002564110621303536,
+ "loss": 4.95107364654541,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.014791073277592659,
+ "learning_rate": 0.002563166827295571,
+ "loss": 4.8995866775512695,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.012700124643743038,
+ "learning_rate": 0.002562222188798525,
+ "loss": 4.922100067138672,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.011235272511839867,
+ "learning_rate": 0.0025612767065734763,
+ "loss": 4.884456634521484,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.013009768910706043,
+ "learning_rate": 0.002560330381382183,
+ "loss": 4.869503021240234,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.0102638378739357,
+ "learning_rate": 0.0025593832139870834,
+ "loss": 4.934551239013672,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.011081870645284653,
+ "learning_rate": 0.0025584352051512913,
+ "loss": 4.937268257141113,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.010681218467652798,
+ "learning_rate": 0.002557486355638601,
+ "loss": 4.920559883117676,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.010962890461087227,
+ "learning_rate": 0.002556536666213484,
+ "loss": 4.936082363128662,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.010790379717946053,
+ "learning_rate": 0.0025555861376410868,
+ "loss": 4.9506449699401855,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.010659157298505306,
+ "learning_rate": 0.0025546347706872346,
+ "loss": 4.872883319854736,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.012962655164301395,
+ "learning_rate": 0.0025536825661184254,
+ "loss": 4.901210784912109,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.014810293912887573,
+ "learning_rate": 0.002552729524701833,
+ "loss": 4.895804405212402,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.010993832722306252,
+ "learning_rate": 0.0025517756472053068,
+ "loss": 4.891035079956055,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.013413915410637856,
+ "learning_rate": 0.002550820934397368,
+ "loss": 4.900317192077637,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.01580147072672844,
+ "learning_rate": 0.0025498653870472117,
+ "loss": 4.883669853210449,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01864512264728546,
+ "learning_rate": 0.0025489090059247045,
+ "loss": 4.881938457489014,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017185743898153305,
+ "learning_rate": 0.0025479517918003862,
+ "loss": 4.853325843811035,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.012045156210660934,
+ "learning_rate": 0.002546993745445466,
+ "loss": 4.9152326583862305,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.013420728035271168,
+ "learning_rate": 0.0025460348676318257,
+ "loss": 4.895130157470703,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.011561610735952854,
+ "learning_rate": 0.0025450751591320152,
+ "loss": 4.890257835388184,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013164008036255836,
+ "learning_rate": 0.0025441146207192547,
+ "loss": 4.885972023010254,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01120204571634531,
+ "learning_rate": 0.002543153253167432,
+ "loss": 4.882246971130371,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.011342627927660942,
+ "learning_rate": 0.002542191057251104,
+ "loss": 4.862927436828613,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.010852213948965073,
+ "learning_rate": 0.0025412280337454942,
+ "loss": 4.879744529724121,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.012851282022893429,
+ "learning_rate": 0.0025402641834264936,
+ "loss": 4.882238388061523,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.011839655227959156,
+ "learning_rate": 0.002539299507070659,
+ "loss": 4.85658073425293,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.011315262876451015,
+ "learning_rate": 0.002538334005455213,
+ "loss": 4.9234771728515625,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.011661105789244175,
+ "learning_rate": 0.0025373676793580416,
+ "loss": 4.855696678161621,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.009712521918118,
+ "learning_rate": 0.0025364005295576975,
+ "loss": 4.89942741394043,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.011002928018569946,
+ "learning_rate": 0.0025354325568333956,
+ "loss": 4.871459007263184,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.01509513333439827,
+ "learning_rate": 0.0025344637619650135,
+ "loss": 4.834778785705566,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 5.031810283660889,
+ "eval_runtime": 51.9412,
+ "eval_samples_per_second": 47.015,
+ "eval_steps_per_second": 1.482,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.010317740961909294,
+ "learning_rate": 0.0025334941457330925,
+ "loss": 4.8169708251953125,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.00892715621739626,
+ "learning_rate": 0.0025325237089188343,
+ "loss": 4.852178573608398,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.011608307249844074,
+ "learning_rate": 0.002531552452304103,
+ "loss": 4.790487289428711,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.013100114651024342,
+ "learning_rate": 0.0025305803766714216,
+ "loss": 4.863536834716797,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.016571057960391045,
+ "learning_rate": 0.002529607482803974,
+ "loss": 4.840987205505371,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.014595622196793556,
+ "learning_rate": 0.002528633771485604,
+ "loss": 4.815027236938477,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.014603931456804276,
+ "learning_rate": 0.002527659243500812,
+ "loss": 4.853024959564209,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.019107041880488396,
+ "learning_rate": 0.0025266838996347595,
+ "loss": 4.829808235168457,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.020292190834879875,
+ "learning_rate": 0.0025257077406732613,
+ "loss": 4.775232315063477,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.014630074612796307,
+ "learning_rate": 0.0025247307674027926,
+ "loss": 4.826682090759277,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.019507668912410736,
+ "learning_rate": 0.0025237529806104823,
+ "loss": 4.843222141265869,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.023950330913066864,
+ "learning_rate": 0.002522774381084116,
+ "loss": 4.820512771606445,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.022547820582985878,
+ "learning_rate": 0.002521794969612133,
+ "loss": 4.854910850524902,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.017819277942180634,
+ "learning_rate": 0.0025208147469836283,
+ "loss": 4.825885772705078,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.017989592626690865,
+ "learning_rate": 0.002519833713988349,
+ "loss": 4.8357014656066895,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.013516640290617943,
+ "learning_rate": 0.0025188518714166956,
+ "loss": 4.83729362487793,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.015956101939082146,
+ "learning_rate": 0.002517869220059721,
+ "loss": 4.807845592498779,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.013967473991215229,
+ "learning_rate": 0.002516885760709131,
+ "loss": 4.828944206237793,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.01270931027829647,
+ "learning_rate": 0.0025159014941572785,
+ "loss": 4.826594352722168,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.012433716095983982,
+ "learning_rate": 0.002514916421197171,
+ "loss": 4.797939777374268,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.011882808059453964,
+ "learning_rate": 0.0025139305426224636,
+ "loss": 4.78917932510376,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.011193005368113518,
+ "learning_rate": 0.0025129438592274614,
+ "loss": 4.787964820861816,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.011873407289385796,
+ "learning_rate": 0.002511956371807116,
+ "loss": 4.81754732131958,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.011407418176531792,
+ "learning_rate": 0.002510968081157029,
+ "loss": 4.781639575958252,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.009560304693877697,
+ "learning_rate": 0.002509978988073449,
+ "loss": 4.8163957595825195,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.009969786740839481,
+ "learning_rate": 0.0025089890933532697,
+ "loss": 4.795045852661133,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.010739767923951149,
+ "learning_rate": 0.0025079983977940312,
+ "loss": 4.847584247589111,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.009999158792197704,
+ "learning_rate": 0.0025070069021939194,
+ "loss": 4.834909439086914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.009302482940256596,
+ "learning_rate": 0.002506014607351765,
+ "loss": 4.808745384216309,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.010687744244933128,
+ "learning_rate": 0.0025050215140670403,
+ "loss": 4.789101600646973,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.010930616408586502,
+ "learning_rate": 0.0025040276231398637,
+ "loss": 4.817464351654053,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.013101065531373024,
+ "learning_rate": 0.0025030329353709956,
+ "loss": 4.827780723571777,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.01223307941108942,
+ "learning_rate": 0.0025020374515618373,
+ "loss": 4.8367414474487305,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.013091634958982468,
+ "learning_rate": 0.0025010411725144323,
+ "loss": 4.839454650878906,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.012114843353629112,
+ "learning_rate": 0.0025000440990314643,
+ "loss": 4.832447052001953,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014664588496088982,
+ "learning_rate": 0.0024990462319162588,
+ "loss": 4.740812301635742,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.015566283836960793,
+ "learning_rate": 0.002498047571972777,
+ "loss": 4.815671920776367,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.012440434657037258,
+ "learning_rate": 0.002497048120005623,
+ "loss": 4.769775867462158,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01175499614328146,
+ "learning_rate": 0.0024960478768200362,
+ "loss": 4.8101348876953125,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.012699508108198643,
+ "learning_rate": 0.002495046843221896,
+ "loss": 4.827281951904297,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.012438303790986538,
+ "learning_rate": 0.002494045020017715,
+ "loss": 4.786289691925049,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.01709924265742302,
+ "learning_rate": 0.002493042408014646,
+ "loss": 4.828742027282715,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.022684535011649132,
+ "learning_rate": 0.0024920390080204733,
+ "loss": 4.822258949279785,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.011984334327280521,
+ "learning_rate": 0.00249103482084362,
+ "loss": 4.792006969451904,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.013941102661192417,
+ "learning_rate": 0.00249002984729314,
+ "loss": 4.7926177978515625,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.0145634226500988,
+ "learning_rate": 0.002489024088178724,
+ "loss": 4.819155216217041,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.013391612097620964,
+ "learning_rate": 0.0024880175443106932,
+ "loss": 4.777242660522461,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01764756441116333,
+ "learning_rate": 0.0024870102165000015,
+ "loss": 4.81627893447876,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.015397441573441029,
+ "learning_rate": 0.002486002105558235,
+ "loss": 4.869495391845703,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01225620973855257,
+ "learning_rate": 0.0024849932122976105,
+ "loss": 4.784395217895508,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.01534759160131216,
+ "learning_rate": 0.002483983537530975,
+ "loss": 4.813475608825684,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.015057779848575592,
+ "learning_rate": 0.002482973082071805,
+ "loss": 4.8281569480896,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.01243986189365387,
+ "learning_rate": 0.0024819618467342066,
+ "loss": 4.7866973876953125,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.013326527550816536,
+ "learning_rate": 0.0024809498323329136,
+ "loss": 4.750214576721191,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.013644695281982422,
+ "learning_rate": 0.002479937039683288,
+ "loss": 4.790144920349121,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01296213548630476,
+ "learning_rate": 0.002478923469601319,
+ "loss": 4.812170028686523,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.011848201975226402,
+ "learning_rate": 0.0024779091229036204,
+ "loss": 4.798855781555176,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.012765968218445778,
+ "learning_rate": 0.0024768940004074335,
+ "loss": 4.8272294998168945,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.011700666509568691,
+ "learning_rate": 0.002475878102930625,
+ "loss": 4.762265205383301,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.01018498931080103,
+ "learning_rate": 0.0024748614312916843,
+ "loss": 4.796628952026367,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.010241681709885597,
+ "learning_rate": 0.002473843986309725,
+ "loss": 4.829319000244141,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.009332080371677876,
+ "learning_rate": 0.0024728257688044855,
+ "loss": 4.783773899078369,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.012860726565122604,
+ "learning_rate": 0.002471806779596324,
+ "loss": 4.802700519561768,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.011156697757542133,
+ "learning_rate": 0.0024707870195062225,
+ "loss": 4.808038234710693,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.010329720564186573,
+ "learning_rate": 0.002469766489355783,
+ "loss": 4.761645317077637,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.011093895882368088,
+ "learning_rate": 0.0024687451899672278,
+ "loss": 4.795220851898193,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.010740782134234905,
+ "learning_rate": 0.0024677231221634,
+ "loss": 4.79289436340332,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.012272508814930916,
+ "learning_rate": 0.0024667002867677606,
+ "loss": 4.811873435974121,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.008665879257023335,
+ "learning_rate": 0.0024656766846043905,
+ "loss": 4.744179725646973,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.00933373998850584,
+ "learning_rate": 0.0024646523164979866,
+ "loss": 4.805453777313232,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.010399337857961655,
+ "learning_rate": 0.002463627183273864,
+ "loss": 4.75568962097168,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.013210986740887165,
+ "learning_rate": 0.0024626012857579542,
+ "loss": 4.797974586486816,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.967026233673096,
+ "eval_runtime": 51.07,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.01731858029961586,
+ "learning_rate": 0.0024615746247768043,
+ "loss": 4.748838901519775,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.018902689218521118,
+ "learning_rate": 0.0024605472011575763,
+ "loss": 4.717338562011719,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.01734415628015995,
+ "learning_rate": 0.0024595190157280468,
+ "loss": 4.737737655639648,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.01971321739256382,
+ "learning_rate": 0.002458490069316607,
+ "loss": 4.774558067321777,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.020029844716191292,
+ "learning_rate": 0.002457460362752259,
+ "loss": 4.739579200744629,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.020853109657764435,
+ "learning_rate": 0.0024564298968646203,
+ "loss": 4.791436672210693,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.019159914925694466,
+ "learning_rate": 0.0024553986724839176,
+ "loss": 4.765683650970459,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.0165473110973835,
+ "learning_rate": 0.0024543666904409903,
+ "loss": 4.732603073120117,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.017426826059818268,
+ "learning_rate": 0.002453333951567287,
+ "loss": 4.7360639572143555,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014590734615921974,
+ "learning_rate": 0.002452300456694868,
+ "loss": 4.750031471252441,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.011172222904860973,
+ "learning_rate": 0.0024512662066563996,
+ "loss": 4.740771293640137,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.012294141575694084,
+ "learning_rate": 0.0024502312022851606,
+ "loss": 4.726214408874512,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.01146282535046339,
+ "learning_rate": 0.002449195444415034,
+ "loss": 4.753599166870117,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.011197247542440891,
+ "learning_rate": 0.0024481589338805113,
+ "loss": 4.751479148864746,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.013252820819616318,
+ "learning_rate": 0.0024471216715166904,
+ "loss": 4.743409156799316,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.012366211041808128,
+ "learning_rate": 0.0024460836581592744,
+ "loss": 4.731625556945801,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.013906655833125114,
+ "learning_rate": 0.0024450448946445737,
+ "loss": 4.768153190612793,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.012118754908442497,
+ "learning_rate": 0.002444005381809499,
+ "loss": 4.74198579788208,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.01197550818324089,
+ "learning_rate": 0.002442965120491568,
+ "loss": 4.737289905548096,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.010640372522175312,
+ "learning_rate": 0.002441924111528901,
+ "loss": 4.7472686767578125,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.013553396798670292,
+ "learning_rate": 0.0024408823557602196,
+ "loss": 4.711215972900391,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.01393711194396019,
+ "learning_rate": 0.002439839854024847,
+ "loss": 4.720060348510742,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01589774526655674,
+ "learning_rate": 0.002438796607162709,
+ "loss": 4.695697784423828,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.0097908740863204,
+ "learning_rate": 0.0024377526160143303,
+ "loss": 4.729333877563477,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.011745489202439785,
+ "learning_rate": 0.0024367078814208363,
+ "loss": 4.722957134246826,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.009649662300944328,
+ "learning_rate": 0.0024356624042239498,
+ "loss": 4.727540493011475,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.008429240435361862,
+ "learning_rate": 0.002434616185265993,
+ "loss": 4.705052375793457,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.00984937883913517,
+ "learning_rate": 0.002433569225389887,
+ "loss": 4.721161365509033,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.009237290360033512,
+ "learning_rate": 0.0024325215254391466,
+ "loss": 4.724277496337891,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.008900545537471771,
+ "learning_rate": 0.002431473086257886,
+ "loss": 4.711522102355957,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.010003026574850082,
+ "learning_rate": 0.002430423908690813,
+ "loss": 4.728816032409668,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.011658324860036373,
+ "learning_rate": 0.0024293739935832316,
+ "loss": 4.716179847717285,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.01312185451388359,
+ "learning_rate": 0.0024283233417810393,
+ "loss": 4.7172346115112305,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.014306646771728992,
+ "learning_rate": 0.0024272719541307266,
+ "loss": 4.704054832458496,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.009844634681940079,
+ "learning_rate": 0.0024262198314793783,
+ "loss": 4.72379207611084,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.00903896801173687,
+ "learning_rate": 0.0024251669746746703,
+ "loss": 4.711025238037109,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.010476433672010899,
+ "learning_rate": 0.0024241133845648703,
+ "loss": 4.711664199829102,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.009128391742706299,
+ "learning_rate": 0.0024230590619988375,
+ "loss": 4.72395133972168,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.01228832732886076,
+ "learning_rate": 0.0024220040078260195,
+ "loss": 4.747347831726074,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.011916602961719036,
+ "learning_rate": 0.002420948222896455,
+ "loss": 4.696471691131592,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.012877700850367546,
+ "learning_rate": 0.0024198917080607706,
+ "loss": 4.731119632720947,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.012967764399945736,
+ "learning_rate": 0.0024188344641701814,
+ "loss": 4.688953876495361,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.011993546038866043,
+ "learning_rate": 0.002417776492076489,
+ "loss": 4.695237159729004,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.011188571341335773,
+ "learning_rate": 0.0024167177926320833,
+ "loss": 4.699850082397461,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.010556348599493504,
+ "learning_rate": 0.0024156583666899395,
+ "loss": 4.703063011169434,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.012023619376122952,
+ "learning_rate": 0.0024145982151036176,
+ "loss": 4.6918792724609375,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01266001258045435,
+ "learning_rate": 0.0024135373387272623,
+ "loss": 4.6653289794921875,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.012152725830674171,
+ "learning_rate": 0.0024124757384156024,
+ "loss": 4.682897567749023,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.014056944288313389,
+ "learning_rate": 0.002411413415023951,
+ "loss": 4.696136474609375,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.016826556995511055,
+ "learning_rate": 0.002410350369408202,
+ "loss": 4.701657295227051,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.015048800967633724,
+ "learning_rate": 0.0024092866024248323,
+ "loss": 4.691389083862305,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.012512968853116035,
+ "learning_rate": 0.0024082221149308997,
+ "loss": 4.677886009216309,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01596425101161003,
+ "learning_rate": 0.0024071569077840425,
+ "loss": 4.718337535858154,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01640196517109871,
+ "learning_rate": 0.002406090981842479,
+ "loss": 4.697303295135498,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.012347170151770115,
+ "learning_rate": 0.002405024337965007,
+ "loss": 4.696938991546631,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01213028747588396,
+ "learning_rate": 0.0024039569770110012,
+ "loss": 4.737306594848633,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01314542256295681,
+ "learning_rate": 0.002402888899840416,
+ "loss": 4.69948673248291,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.013764189556241035,
+ "learning_rate": 0.0024018201073137805,
+ "loss": 4.722800254821777,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.011147351004183292,
+ "learning_rate": 0.002400750600292203,
+ "loss": 4.694892883300781,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.015437614172697067,
+ "learning_rate": 0.0023996803796373653,
+ "loss": 4.704752445220947,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.016402097418904305,
+ "learning_rate": 0.0023986094462115255,
+ "loss": 4.67849063873291,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.011600594967603683,
+ "learning_rate": 0.002397537800877514,
+ "loss": 4.689641952514648,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.011864298023283482,
+ "learning_rate": 0.0023964654444987373,
+ "loss": 4.670377731323242,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01409872155636549,
+ "learning_rate": 0.002395392377939173,
+ "loss": 4.696022033691406,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.012106554582715034,
+ "learning_rate": 0.0023943186020633717,
+ "loss": 4.705615043640137,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.012310667894780636,
+ "learning_rate": 0.002393244117736455,
+ "loss": 4.702986717224121,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.013501073233783245,
+ "learning_rate": 0.002392168925824116,
+ "loss": 4.6700215339660645,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.013565126806497574,
+ "learning_rate": 0.0023910930271926166,
+ "loss": 4.699914455413818,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.013421077281236649,
+ "learning_rate": 0.00239001642270879,
+ "loss": 4.713191986083984,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.014408102259039879,
+ "learning_rate": 0.002388939113240035,
+ "loss": 4.655015468597412,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.013502835296094418,
+ "learning_rate": 0.0023878610996543224,
+ "loss": 4.6142258644104,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.014169570058584213,
+ "learning_rate": 0.002386782382820188,
+ "loss": 4.660243034362793,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.885881423950195,
+ "eval_runtime": 52.0315,
+ "eval_samples_per_second": 46.933,
+ "eval_steps_per_second": 1.48,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.014618083834648132,
+ "learning_rate": 0.002385702963606734,
+ "loss": 4.640046119689941,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01745622046291828,
+ "learning_rate": 0.0023846228428836278,
+ "loss": 4.641966819763184,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.014995258301496506,
+ "learning_rate": 0.002383542021521106,
+ "loss": 4.60922908782959,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.013278020545840263,
+ "learning_rate": 0.002382460500389964,
+ "loss": 4.652071952819824,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.014685257337987423,
+ "learning_rate": 0.002381378280361566,
+ "loss": 4.672174453735352,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.012737024575471878,
+ "learning_rate": 0.002380295362307836,
+ "loss": 4.612064361572266,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.012992842122912407,
+ "learning_rate": 0.002379211747101261,
+ "loss": 4.611995697021484,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.010210646316409111,
+ "learning_rate": 0.0023781274356148912,
+ "loss": 4.6055707931518555,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.010265499353408813,
+ "learning_rate": 0.0023770424287223365,
+ "loss": 4.637089729309082,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.010174009948968887,
+ "learning_rate": 0.002375956727297768,
+ "loss": 4.620571136474609,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.009512446820735931,
+ "learning_rate": 0.0023748703322159143,
+ "loss": 4.644092559814453,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.009849252179265022,
+ "learning_rate": 0.0023737832443520655,
+ "loss": 4.625055313110352,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.010493188165128231,
+ "learning_rate": 0.0023726954645820674,
+ "loss": 4.644410610198975,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.022181199863553047,
+ "learning_rate": 0.002371606993782326,
+ "loss": 4.593084335327148,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.011906919069588184,
+ "learning_rate": 0.0023705178328298016,
+ "loss": 4.6452836990356445,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.011874495074152946,
+ "learning_rate": 0.0023694279826020122,
+ "loss": 4.599322319030762,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.00989708211272955,
+ "learning_rate": 0.0023683374439770295,
+ "loss": 4.607342720031738,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.009256849065423012,
+ "learning_rate": 0.0023672462178334817,
+ "loss": 4.621583461761475,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.013187132775783539,
+ "learning_rate": 0.0023661543050505496,
+ "loss": 4.614629745483398,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.013505950570106506,
+ "learning_rate": 0.002365061706507968,
+ "loss": 4.629693984985352,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.012032733298838139,
+ "learning_rate": 0.0023639684230860225,
+ "loss": 4.671209335327148,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.01278337650001049,
+ "learning_rate": 0.002362874455665553,
+ "loss": 4.648874759674072,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.01123690977692604,
+ "learning_rate": 0.0023617798051279496,
+ "loss": 4.611470699310303,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.012564311735332012,
+ "learning_rate": 0.002360684472355151,
+ "loss": 4.659026145935059,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.01206638477742672,
+ "learning_rate": 0.002359588458229649,
+ "loss": 4.642566204071045,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.010300020687282085,
+ "learning_rate": 0.0023584917636344806,
+ "loss": 4.629585266113281,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.0102265365421772,
+ "learning_rate": 0.0023573943894532336,
+ "loss": 4.609250068664551,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.012440572492778301,
+ "learning_rate": 0.002356296336570042,
+ "loss": 4.636714458465576,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.011830241419374943,
+ "learning_rate": 0.0023551976058695887,
+ "loss": 4.59100866317749,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.011482699774205685,
+ "learning_rate": 0.0023540981982370997,
+ "loss": 4.620088577270508,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.012950294651091099,
+ "learning_rate": 0.0023529981145583483,
+ "loss": 4.578657627105713,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.012197710573673248,
+ "learning_rate": 0.002351897355719652,
+ "loss": 4.620813846588135,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.014839230105280876,
+ "learning_rate": 0.002350795922607872,
+ "loss": 4.615731239318848,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016670791432261467,
+ "learning_rate": 0.002349693816110414,
+ "loss": 4.655855178833008,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.014426100067794323,
+ "learning_rate": 0.0023485910371152247,
+ "loss": 4.660849571228027,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.014040560461580753,
+ "learning_rate": 0.0023474875865107936,
+ "loss": 4.634917736053467,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018269391730427742,
+ "learning_rate": 0.0023463834651861496,
+ "loss": 4.611809253692627,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.013855630531907082,
+ "learning_rate": 0.0023452786740308647,
+ "loss": 4.615225791931152,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.011505477130413055,
+ "learning_rate": 0.0023441732139350487,
+ "loss": 4.62592887878418,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.010166114196181297,
+ "learning_rate": 0.002343067085789351,
+ "loss": 4.6219024658203125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.009253021329641342,
+ "learning_rate": 0.0023419602904849596,
+ "loss": 4.658320426940918,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01027526892721653,
+ "learning_rate": 0.002340852828913598,
+ "loss": 4.589777946472168,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.008855022490024567,
+ "learning_rate": 0.002339744701967529,
+ "loss": 4.588204383850098,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.00965463649481535,
+ "learning_rate": 0.002338635910539551,
+ "loss": 4.648177146911621,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.009073720313608646,
+ "learning_rate": 0.0023375264555229956,
+ "loss": 4.585582733154297,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.010105383582413197,
+ "learning_rate": 0.0023364163378117317,
+ "loss": 4.612174987792969,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.015307556837797165,
+ "learning_rate": 0.002335305558300161,
+ "loss": 4.591984748840332,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016716254875063896,
+ "learning_rate": 0.002334194117883219,
+ "loss": 4.659285545349121,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.013281598687171936,
+ "learning_rate": 0.0023330820174563725,
+ "loss": 4.645829200744629,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.014416513964533806,
+ "learning_rate": 0.0023319692579156204,
+ "loss": 4.584893226623535,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.016326071694493294,
+ "learning_rate": 0.002330855840157494,
+ "loss": 4.633223533630371,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.012851824052631855,
+ "learning_rate": 0.002329741765079053,
+ "loss": 4.580986499786377,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.015444506891071796,
+ "learning_rate": 0.0023286270335778876,
+ "loss": 4.615891456604004,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.01391979493200779,
+ "learning_rate": 0.0023275116465521174,
+ "loss": 4.628403186798096,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.012138686142861843,
+ "learning_rate": 0.0023263956049003895,
+ "loss": 4.6259942054748535,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.013741672039031982,
+ "learning_rate": 0.0023252789095218785,
+ "loss": 4.6056742668151855,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01103319600224495,
+ "learning_rate": 0.002324161561316285,
+ "loss": 4.621670246124268,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.010333316400647163,
+ "learning_rate": 0.0023230435611838367,
+ "loss": 4.620484352111816,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.011063606478273869,
+ "learning_rate": 0.0023219249100252864,
+ "loss": 4.64952278137207,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.009984857402741909,
+ "learning_rate": 0.0023208056087419106,
+ "loss": 4.572571754455566,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.009986856952309608,
+ "learning_rate": 0.002319685658235511,
+ "loss": 4.570366382598877,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.010905452072620392,
+ "learning_rate": 0.0023185650594084103,
+ "loss": 4.603614807128906,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.012800061143934727,
+ "learning_rate": 0.002317443813163456,
+ "loss": 4.562705039978027,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.01252608560025692,
+ "learning_rate": 0.002316321920404015,
+ "loss": 4.599016189575195,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.011305592954158783,
+ "learning_rate": 0.002315199382033976,
+ "loss": 4.644617557525635,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.010594413615763187,
+ "learning_rate": 0.002314076198957749,
+ "loss": 4.569808006286621,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.009859138168394566,
+ "learning_rate": 0.002312952372080261,
+ "loss": 4.5890398025512695,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.010982917621731758,
+ "learning_rate": 0.00231182790230696,
+ "loss": 4.606348991394043,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.010252755135297775,
+ "learning_rate": 0.00231070279054381,
+ "loss": 4.58942174911499,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01110257301479578,
+ "learning_rate": 0.0023095770376972947,
+ "loss": 4.599823474884033,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01003815233707428,
+ "learning_rate": 0.0023084506446744103,
+ "loss": 4.569343090057373,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.010517258197069168,
+ "learning_rate": 0.0023073236123826745,
+ "loss": 4.646301746368408,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.804611682891846,
+ "eval_runtime": 51.6279,
+ "eval_samples_per_second": 47.3,
+ "eval_steps_per_second": 1.491,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.011439066380262375,
+ "learning_rate": 0.0023061959417301145,
+ "loss": 4.563335418701172,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.01073157787322998,
+ "learning_rate": 0.002305067633625275,
+ "loss": 4.540966987609863,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.011391228064894676,
+ "learning_rate": 0.0023039386889772136,
+ "loss": 4.555583953857422,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.012814595364034176,
+ "learning_rate": 0.0023028091086955,
+ "loss": 4.517425537109375,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.012994988821446896,
+ "learning_rate": 0.002301678893690218,
+ "loss": 4.538477897644043,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.01460991520434618,
+ "learning_rate": 0.00230054804487196,
+ "loss": 4.571145057678223,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.013036075979471207,
+ "learning_rate": 0.0022994165631518306,
+ "loss": 4.519233703613281,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.011917057447135448,
+ "learning_rate": 0.0022982844494414444,
+ "loss": 4.5071587562561035,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.013928866013884544,
+ "learning_rate": 0.0022971517046529257,
+ "loss": 4.531422138214111,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01407038513571024,
+ "learning_rate": 0.0022960183296989053,
+ "loss": 4.511205196380615,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.019396567717194557,
+ "learning_rate": 0.0022948843254925245,
+ "loss": 4.530007362365723,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.014425835572183132,
+ "learning_rate": 0.002293749692947428,
+ "loss": 4.530138969421387,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.01600247249007225,
+ "learning_rate": 0.00229261443297777,
+ "loss": 4.5643391609191895,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.018238037824630737,
+ "learning_rate": 0.0022914785464982095,
+ "loss": 4.540779113769531,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.015427535399794579,
+ "learning_rate": 0.0022903420344239085,
+ "loss": 4.506287097930908,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.016816625371575356,
+ "learning_rate": 0.0022892048976705347,
+ "loss": 4.537919521331787,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.01999717950820923,
+ "learning_rate": 0.002288067137154259,
+ "loss": 4.558426380157471,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.020884918048977852,
+ "learning_rate": 0.002286928753791753,
+ "loss": 4.592915058135986,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.01620141603052616,
+ "learning_rate": 0.002285789748500194,
+ "loss": 4.553356170654297,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.018495604395866394,
+ "learning_rate": 0.0022846501221972567,
+ "loss": 4.573139190673828,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.0156320258975029,
+ "learning_rate": 0.0022835098758011174,
+ "loss": 4.585921764373779,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.014515317976474762,
+ "learning_rate": 0.002282369010230452,
+ "loss": 4.5532002449035645,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.014826556667685509,
+ "learning_rate": 0.0022812275264044355,
+ "loss": 4.560998916625977,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.013240831904113293,
+ "learning_rate": 0.00228008542524274,
+ "loss": 4.531787872314453,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.013814719393849373,
+ "learning_rate": 0.0022789427076655372,
+ "loss": 4.611872673034668,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.012246561236679554,
+ "learning_rate": 0.0022777993745934934,
+ "loss": 4.5565385818481445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01206218171864748,
+ "learning_rate": 0.002276655426947771,
+ "loss": 4.570852279663086,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01007329672574997,
+ "learning_rate": 0.0022755108656500275,
+ "loss": 4.553400039672852,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.010361720807850361,
+ "learning_rate": 0.0022743656916224173,
+ "loss": 4.527509689331055,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.009456256404519081,
+ "learning_rate": 0.0022732199057875846,
+ "loss": 4.520236015319824,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.009310049936175346,
+ "learning_rate": 0.002272073509068669,
+ "loss": 4.518343925476074,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.00927449855953455,
+ "learning_rate": 0.0022709265023893015,
+ "loss": 4.559690475463867,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.008658594451844692,
+ "learning_rate": 0.002269778886673605,
+ "loss": 4.513876914978027,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.008928220719099045,
+ "learning_rate": 0.0022686306628461918,
+ "loss": 4.545050621032715,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.008046358823776245,
+ "learning_rate": 0.002267481831832166,
+ "loss": 4.54866361618042,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.008318924345076084,
+ "learning_rate": 0.00226633239455712,
+ "loss": 4.546968936920166,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.008808869868516922,
+ "learning_rate": 0.0022651823519471343,
+ "loss": 4.50877571105957,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.008767383173108101,
+ "learning_rate": 0.0022640317049287774,
+ "loss": 4.57666540145874,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.009351354092359543,
+ "learning_rate": 0.0022628804544291047,
+ "loss": 4.511702537536621,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.010137057863175869,
+ "learning_rate": 0.002261728601375659,
+ "loss": 4.539676189422607,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.00882046204060316,
+ "learning_rate": 0.002260576146696466,
+ "loss": 4.533196449279785,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.009354782290756702,
+ "learning_rate": 0.0022594230913200388,
+ "loss": 4.515488624572754,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.009241201914846897,
+ "learning_rate": 0.002258269436175372,
+ "loss": 4.542706489562988,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.009252829477190971,
+ "learning_rate": 0.002257115182191947,
+ "loss": 4.592660903930664,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.011243131943047047,
+ "learning_rate": 0.0022559603302997225,
+ "loss": 4.527804851531982,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.011032823473215103,
+ "learning_rate": 0.0022548048814291437,
+ "loss": 4.540122032165527,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.010495275259017944,
+ "learning_rate": 0.0022536488365111347,
+ "loss": 4.532835006713867,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.009992090053856373,
+ "learning_rate": 0.0022524921964770994,
+ "loss": 4.521217346191406,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01141198631376028,
+ "learning_rate": 0.0022513349622589225,
+ "loss": 4.519709587097168,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.013087889179587364,
+ "learning_rate": 0.002250177134788966,
+ "loss": 4.547289848327637,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.01328477542847395,
+ "learning_rate": 0.0022490187150000706,
+ "loss": 4.527865409851074,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01033794041723013,
+ "learning_rate": 0.002247859703825555,
+ "loss": 4.54740047454834,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.010522300377488136,
+ "learning_rate": 0.002246700102199213,
+ "loss": 4.538840293884277,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.012151090428233147,
+ "learning_rate": 0.0022455399110553145,
+ "loss": 4.536865234375,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.010553102008998394,
+ "learning_rate": 0.0022443791313286054,
+ "loss": 4.5348334312438965,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.011433709412813187,
+ "learning_rate": 0.0022432177639543035,
+ "loss": 4.510526657104492,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01189691387116909,
+ "learning_rate": 0.002242055809868103,
+ "loss": 4.530807018280029,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.010031254962086678,
+ "learning_rate": 0.0022408932700061683,
+ "loss": 4.496850967407227,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.010129331611096859,
+ "learning_rate": 0.002239730145305138,
+ "loss": 4.551287651062012,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.010430586524307728,
+ "learning_rate": 0.0022385664367021185,
+ "loss": 4.542672157287598,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.011774476617574692,
+ "learning_rate": 0.002237402145134691,
+ "loss": 4.537034034729004,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.01184659544378519,
+ "learning_rate": 0.002236237271540903,
+ "loss": 4.521193504333496,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.011327247135341167,
+ "learning_rate": 0.0022350718168592724,
+ "loss": 4.49426794052124,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.010767918080091476,
+ "learning_rate": 0.0022339057820287855,
+ "loss": 4.531225204467773,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.011195424944162369,
+ "learning_rate": 0.002232739167988894,
+ "loss": 4.492671012878418,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.010394908487796783,
+ "learning_rate": 0.0022315719756795197,
+ "loss": 4.523616790771484,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.010277998633682728,
+ "learning_rate": 0.002230404206041047,
+ "loss": 4.487832069396973,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.010955783538520336,
+ "learning_rate": 0.0022292358600143275,
+ "loss": 4.526865005493164,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.010611076839268208,
+ "learning_rate": 0.002228066938540676,
+ "loss": 4.498682022094727,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.010008828714489937,
+ "learning_rate": 0.0022268974425618724,
+ "loss": 4.524019241333008,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.008973540738224983,
+ "learning_rate": 0.002225727373020157,
+ "loss": 4.512996673583984,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.011754963546991348,
+ "learning_rate": 0.002224556730858235,
+ "loss": 4.496376991271973,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.744190216064453,
+ "eval_runtime": 51.6572,
+ "eval_samples_per_second": 47.273,
+ "eval_steps_per_second": 1.491,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.012094628997147083,
+ "learning_rate": 0.0022233855170192703,
+ "loss": 4.443447113037109,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.014347990974783897,
+ "learning_rate": 0.00222221373244689,
+ "loss": 4.485690116882324,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.021286984905600548,
+ "learning_rate": 0.0022210413780851786,
+ "loss": 4.487505912780762,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.020273910835385323,
+ "learning_rate": 0.0022198684548786827,
+ "loss": 4.473119735717773,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.017642971128225327,
+ "learning_rate": 0.002218694963772403,
+ "loss": 4.440615653991699,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.02075147069990635,
+ "learning_rate": 0.0022175209057118017,
+ "loss": 4.527026176452637,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016479508951306343,
+ "learning_rate": 0.002216346281642796,
+ "loss": 4.508321762084961,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.017978593707084656,
+ "learning_rate": 0.002215171092511759,
+ "loss": 4.500985145568848,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.01739897020161152,
+ "learning_rate": 0.002213995339265519,
+ "loss": 4.474225044250488,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.015556776896119118,
+ "learning_rate": 0.0022128190228513597,
+ "loss": 4.444849014282227,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.016166578978300095,
+ "learning_rate": 0.002211642144217018,
+ "loss": 4.5026726722717285,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.01534261740744114,
+ "learning_rate": 0.0022104647043106832,
+ "loss": 4.484527587890625,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.013517106883227825,
+ "learning_rate": 0.002209286704080998,
+ "loss": 4.453907012939453,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.012562787160277367,
+ "learning_rate": 0.0022081081444770556,
+ "loss": 4.471528053283691,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.009947682730853558,
+ "learning_rate": 0.0022069290264483997,
+ "loss": 4.504153728485107,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01169571466743946,
+ "learning_rate": 0.002205749350945026,
+ "loss": 4.447345733642578,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.011394352652132511,
+ "learning_rate": 0.0022045691189173755,
+ "loss": 4.447491645812988,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.011168593540787697,
+ "learning_rate": 0.0022033883313163417,
+ "loss": 4.441639423370361,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.009503341279923916,
+ "learning_rate": 0.002202206989093263,
+ "loss": 4.461601257324219,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.0112199941650033,
+ "learning_rate": 0.002201025093199926,
+ "loss": 4.444095134735107,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.010505333542823792,
+ "learning_rate": 0.002199842644588562,
+ "loss": 4.460987091064453,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.009494813159108162,
+ "learning_rate": 0.0021986596442118496,
+ "loss": 4.460158348083496,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.01063452661037445,
+ "learning_rate": 0.00219747609302291,
+ "loss": 4.489625453948975,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.012194888666272163,
+ "learning_rate": 0.0021962919919753087,
+ "loss": 4.441605567932129,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.010613647289574146,
+ "learning_rate": 0.0021951073420230557,
+ "loss": 4.475821018218994,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.009643700905144215,
+ "learning_rate": 0.002193922144120601,
+ "loss": 4.467228889465332,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.00912613607943058,
+ "learning_rate": 0.0021927363992228373,
+ "loss": 4.455435752868652,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.009371735155582428,
+ "learning_rate": 0.0021915501082850988,
+ "loss": 4.479915142059326,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.008617875166237354,
+ "learning_rate": 0.002190363272263158,
+ "loss": 4.415642738342285,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.009571383707225323,
+ "learning_rate": 0.002189175892113227,
+ "loss": 4.439854621887207,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.00782635249197483,
+ "learning_rate": 0.002187987968791957,
+ "loss": 4.484311103820801,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.008408057503402233,
+ "learning_rate": 0.0021867995032564358,
+ "loss": 4.46577262878418,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.009641910903155804,
+ "learning_rate": 0.00218561049646419,
+ "loss": 4.437408447265625,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.008498182520270348,
+ "learning_rate": 0.0021844209493731784,
+ "loss": 4.442599296569824,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.009289798326790333,
+ "learning_rate": 0.0021832308629418,
+ "loss": 4.517641067504883,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.009360183961689472,
+ "learning_rate": 0.0021820402381288848,
+ "loss": 4.474143981933594,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.008764738216996193,
+ "learning_rate": 0.0021808490758936983,
+ "loss": 4.4376702308654785,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.009673701599240303,
+ "learning_rate": 0.0021796573771959375,
+ "loss": 4.435220718383789,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.010852573439478874,
+ "learning_rate": 0.0021784651429957334,
+ "loss": 4.465187072753906,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.011292314156889915,
+ "learning_rate": 0.0021772723742536464,
+ "loss": 4.458176136016846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01071334257721901,
+ "learning_rate": 0.0021760790719306702,
+ "loss": 4.425498962402344,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.01505239401012659,
+ "learning_rate": 0.0021748852369882265,
+ "loss": 4.463129043579102,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.011774689890444279,
+ "learning_rate": 0.0021736908703881663,
+ "loss": 4.462577819824219,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.011241118423640728,
+ "learning_rate": 0.002172495973092769,
+ "loss": 4.452583312988281,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.01072910986840725,
+ "learning_rate": 0.002171300546064742,
+ "loss": 4.466248035430908,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01021308358758688,
+ "learning_rate": 0.002170104590267219,
+ "loss": 4.485857963562012,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.012630444951355457,
+ "learning_rate": 0.00216890810666376,
+ "loss": 4.442112922668457,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.013602569699287415,
+ "learning_rate": 0.0021677110962183515,
+ "loss": 4.45171594619751,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.013808641582727432,
+ "learning_rate": 0.0021665135598954014,
+ "loss": 4.468818187713623,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.011510699987411499,
+ "learning_rate": 0.002165315498659744,
+ "loss": 4.439584255218506,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.010502424091100693,
+ "learning_rate": 0.0021641169134766353,
+ "loss": 4.478377342224121,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.009953407570719719,
+ "learning_rate": 0.0021629178053117542,
+ "loss": 4.474404335021973,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.011158930137753487,
+ "learning_rate": 0.0021617181751312,
+ "loss": 4.4549560546875,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.010431940667331219,
+ "learning_rate": 0.002160518023901493,
+ "loss": 4.444911003112793,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.009100335650146008,
+ "learning_rate": 0.0021593173525895734,
+ "loss": 4.428071975708008,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.009362847544252872,
+ "learning_rate": 0.0021581161621628013,
+ "loss": 4.462827682495117,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.009893129579722881,
+ "learning_rate": 0.0021569144535889535,
+ "loss": 4.426912307739258,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.010524468496441841,
+ "learning_rate": 0.0021557122278362244,
+ "loss": 4.435606002807617,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.010621445253491402,
+ "learning_rate": 0.002154509485873226,
+ "loss": 4.454901695251465,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.010836896486580372,
+ "learning_rate": 0.0021533062286689855,
+ "loss": 4.4441728591918945,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.011403030715882778,
+ "learning_rate": 0.0021521024571929457,
+ "loss": 4.483884811401367,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.012522426433861256,
+ "learning_rate": 0.0021508981724149643,
+ "loss": 4.464006423950195,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.014799179509282112,
+ "learning_rate": 0.0021496933753053105,
+ "loss": 4.460402011871338,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.012453878298401833,
+ "learning_rate": 0.0021484880668346685,
+ "loss": 4.456899642944336,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.012270379811525345,
+ "learning_rate": 0.002147282247974133,
+ "loss": 4.477325439453125,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01094560045748949,
+ "learning_rate": 0.0021460759196952105,
+ "loss": 4.422362327575684,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.009995176456868649,
+ "learning_rate": 0.0021448690829698183,
+ "loss": 4.405127048492432,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.014935536310076714,
+ "learning_rate": 0.002143661738770283,
+ "loss": 4.445840358734131,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.011492030695080757,
+ "learning_rate": 0.0021424538880693393,
+ "loss": 4.459555625915527,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.008793608285486698,
+ "learning_rate": 0.002141245531840132,
+ "loss": 4.436631202697754,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.008894025348126888,
+ "learning_rate": 0.0021400366710562103,
+ "loss": 4.473924160003662,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.010324863716959953,
+ "learning_rate": 0.0021388273066915327,
+ "loss": 4.475680351257324,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.688324451446533,
+ "eval_runtime": 54.4515,
+ "eval_samples_per_second": 44.847,
+ "eval_steps_per_second": 1.414,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.009235628880560398,
+ "learning_rate": 0.0021376174397204615,
+ "loss": 4.370064735412598,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.008818200789391994,
+ "learning_rate": 0.0021364070711177643,
+ "loss": 4.371678352355957,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.010333757847547531,
+ "learning_rate": 0.0021351962018586147,
+ "loss": 4.3896379470825195,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.011938809417188168,
+ "learning_rate": 0.002133984832918587,
+ "loss": 4.361343860626221,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.010634614154696465,
+ "learning_rate": 0.00213277296527366,
+ "loss": 4.362449645996094,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.008545687422156334,
+ "learning_rate": 0.0021315605999002126,
+ "loss": 4.3685503005981445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.009159795008599758,
+ "learning_rate": 0.002130347737775027,
+ "loss": 4.379969596862793,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.010464153252542019,
+ "learning_rate": 0.0021291343798752835,
+ "loss": 4.376113414764404,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.011010527610778809,
+ "learning_rate": 0.0021279205271785622,
+ "loss": 4.390711784362793,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.010350867174565792,
+ "learning_rate": 0.002126706180662843,
+ "loss": 4.399683952331543,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.009708237834274769,
+ "learning_rate": 0.002125491341306503,
+ "loss": 4.363435745239258,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.010087115690112114,
+ "learning_rate": 0.002124276010088316,
+ "loss": 4.366623878479004,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.009712349623441696,
+ "learning_rate": 0.002123060187987453,
+ "loss": 4.360814571380615,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.010174986906349659,
+ "learning_rate": 0.0021218438759834785,
+ "loss": 4.347475051879883,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.010195785202085972,
+ "learning_rate": 0.0021206270750563543,
+ "loss": 4.399251461029053,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.010786959901452065,
+ "learning_rate": 0.0021194097861864354,
+ "loss": 4.397933006286621,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.011476189829409122,
+ "learning_rate": 0.002118192010354468,
+ "loss": 4.3748369216918945,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.011571013368666172,
+ "learning_rate": 0.0021169737485415927,
+ "loss": 4.388349533081055,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.011140935122966766,
+ "learning_rate": 0.002115755001729341,
+ "loss": 4.398018836975098,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.013216734863817692,
+ "learning_rate": 0.0021145357708996365,
+ "loss": 4.406947135925293,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.013385126367211342,
+ "learning_rate": 0.0021133160570347893,
+ "loss": 4.41457462310791,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.011784231290221214,
+ "learning_rate": 0.002112095861117502,
+ "loss": 4.403835296630859,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.009813498705625534,
+ "learning_rate": 0.0021108751841308645,
+ "loss": 4.3865647315979,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.010297936387360096,
+ "learning_rate": 0.0021096540270583533,
+ "loss": 4.395583152770996,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.011965147219598293,
+ "learning_rate": 0.002108432390883834,
+ "loss": 4.385443687438965,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.012212193571031094,
+ "learning_rate": 0.0021072102765915556,
+ "loss": 4.422321319580078,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.009840975515544415,
+ "learning_rate": 0.0021059876851661544,
+ "loss": 4.3568902015686035,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.013039806857705116,
+ "learning_rate": 0.002104764617592649,
+ "loss": 4.314867973327637,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.015091902576386929,
+ "learning_rate": 0.0021035410748564444,
+ "loss": 4.392338752746582,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.015448705293238163,
+ "learning_rate": 0.0021023170579433257,
+ "loss": 4.379609107971191,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.018393104895949364,
+ "learning_rate": 0.0021010925678394622,
+ "loss": 4.418291091918945,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.01866372860968113,
+ "learning_rate": 0.002099867605531402,
+ "loss": 4.4006452560424805,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.01819106750190258,
+ "learning_rate": 0.002098642172006077,
+ "loss": 4.433975696563721,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.019774602726101875,
+ "learning_rate": 0.0020974162682507952,
+ "loss": 4.417821884155273,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.016885802149772644,
+ "learning_rate": 0.0020961898952532457,
+ "loss": 4.397073268890381,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.015750881284475327,
+ "learning_rate": 0.0020949630540014956,
+ "loss": 4.4356608390808105,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.015696514397859573,
+ "learning_rate": 0.002093735745483988,
+ "loss": 4.422449588775635,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.017195021733641624,
+ "learning_rate": 0.0020925079706895435,
+ "loss": 4.443209648132324,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.015864048153162003,
+ "learning_rate": 0.0020912797306073577,
+ "loss": 4.4340691566467285,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.009857800789177418,
+ "learning_rate": 0.0020900510262270022,
+ "loss": 4.37935209274292,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.013739476911723614,
+ "learning_rate": 0.0020888218585384216,
+ "loss": 4.412148475646973,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.011667690239846706,
+ "learning_rate": 0.0020875922285319344,
+ "loss": 4.391700744628906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.012323460541665554,
+ "learning_rate": 0.0020863621371982302,
+ "loss": 4.423880577087402,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.010353626683354378,
+ "learning_rate": 0.0020851315855283724,
+ "loss": 4.427382469177246,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.0101751908659935,
+ "learning_rate": 0.002083900574513794,
+ "loss": 4.415317535400391,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.01126549206674099,
+ "learning_rate": 0.002082669105146298,
+ "loss": 4.402498245239258,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.011096738278865814,
+ "learning_rate": 0.0020814371784180575,
+ "loss": 4.433567047119141,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.010944445617496967,
+ "learning_rate": 0.0020802047953216137,
+ "loss": 4.406990051269531,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.008965318091213703,
+ "learning_rate": 0.0020789719568498747,
+ "loss": 4.3721771240234375,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.00828113779425621,
+ "learning_rate": 0.002077738663996117,
+ "loss": 4.389828205108643,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.008599454537034035,
+ "learning_rate": 0.002076504917753982,
+ "loss": 4.411396026611328,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.007713889703154564,
+ "learning_rate": 0.0020752707191174762,
+ "loss": 4.355343341827393,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.00912355538457632,
+ "learning_rate": 0.0020740360690809723,
+ "loss": 4.423413276672363,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.008269775658845901,
+ "learning_rate": 0.0020728009686392045,
+ "loss": 4.390920639038086,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.007461601868271828,
+ "learning_rate": 0.002071565418787272,
+ "loss": 4.365501403808594,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.007951333187520504,
+ "learning_rate": 0.0020703294205206336,
+ "loss": 4.378884315490723,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.007816503755748272,
+ "learning_rate": 0.002069092974835112,
+ "loss": 4.4241743087768555,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.00833756010979414,
+ "learning_rate": 0.0020678560827268886,
+ "loss": 4.391102313995361,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.008241684176027775,
+ "learning_rate": 0.0020666187451925055,
+ "loss": 4.374970436096191,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.008245996199548244,
+ "learning_rate": 0.002065380963228862,
+ "loss": 4.380222320556641,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.007378706708550453,
+ "learning_rate": 0.0020641427378332183,
+ "loss": 4.391451358795166,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.007594958413392305,
+ "learning_rate": 0.002062904070003189,
+ "loss": 4.372774124145508,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.0076752156019210815,
+ "learning_rate": 0.002061664960736747,
+ "loss": 4.384668827056885,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.008727089501917362,
+ "learning_rate": 0.00206042541103222,
+ "loss": 4.363401412963867,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.007616396062076092,
+ "learning_rate": 0.002059185421888291,
+ "loss": 4.413188934326172,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.008922954089939594,
+ "learning_rate": 0.0020579449943039967,
+ "loss": 4.383417129516602,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.009063739329576492,
+ "learning_rate": 0.002056704129278727,
+ "loss": 4.385283470153809,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.008912574499845505,
+ "learning_rate": 0.002055462827812225,
+ "loss": 4.378478050231934,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.008317525498569012,
+ "learning_rate": 0.0020542210909045844,
+ "loss": 4.367859840393066,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.009347227402031422,
+ "learning_rate": 0.0020529789195562506,
+ "loss": 4.38238000869751,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.011030065827071667,
+ "learning_rate": 0.0020517363147680183,
+ "loss": 4.372624397277832,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.011519700288772583,
+ "learning_rate": 0.002050493277541032,
+ "loss": 4.3906121253967285,
+ "step": 1440
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 6.139226722664448e+17,
+ "train_batch_size": 16,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/baseline/checkpoint-1440/training_args.bin b/runs/baseline/checkpoint-1440/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7
--- /dev/null
+++ b/runs/baseline/checkpoint-1440/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97
+size 5265
diff --git a/runs/baseline/checkpoint-1800/chat_template.jinja b/runs/baseline/checkpoint-1800/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/baseline/checkpoint-1800/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/baseline/checkpoint-1800/config.json b/runs/baseline/checkpoint-1800/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/baseline/checkpoint-1800/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/baseline/checkpoint-1800/generation_config.json b/runs/baseline/checkpoint-1800/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/baseline/checkpoint-1800/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/baseline/checkpoint-1800/model.safetensors b/runs/baseline/checkpoint-1800/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..8327df03d22e26408ccf723ec1ca3f534538fa36
--- /dev/null
+++ b/runs/baseline/checkpoint-1800/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0c2ac76aaab232d5b01beedf4207ba5eea5d85fe47eae4f88f5cf93160abb37d
+size 583356232
diff --git a/runs/baseline/checkpoint-1800/optimizer.pt b/runs/baseline/checkpoint-1800/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..7448dcb39d76879aae6cff91ec20c9f64cbd77a0
--- /dev/null
+++ b/runs/baseline/checkpoint-1800/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:13614eed9a3624e21a4f90498242d9bb1c4283ed2a4e2786e604d0a083b1cb5a
+size 1166825803
diff --git a/runs/baseline/checkpoint-1800/rng_state_0.pth b/runs/baseline/checkpoint-1800/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..304a7576d9bd060137258dba420a3e1ff00587c1
--- /dev/null
+++ b/runs/baseline/checkpoint-1800/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:987694c44f0af3a8ae4e43e3677bef48a1cfa8e57eee3832ef223e9b8ad80516
+size 14917
diff --git a/runs/baseline/checkpoint-1800/rng_state_1.pth b/runs/baseline/checkpoint-1800/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..75486c47bfa9499bfe91975e3af3355c9526365c
--- /dev/null
+++ b/runs/baseline/checkpoint-1800/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ea5ecba78b648c2a45cab91d590e52b7295077372e6bf91c1591d3fb6a5535ca
+size 14917
diff --git a/runs/baseline/checkpoint-1800/scheduler.pt b/runs/baseline/checkpoint-1800/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..7bf0bbb2d908712e094e342325183f9d764f7a64
--- /dev/null
+++ b/runs/baseline/checkpoint-1800/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c1662fc1ce9587c318535917c43a6b1f1e5532c3833011159b3431966308a9b2
+size 1465
diff --git a/runs/baseline/checkpoint-1800/tokenizer.json b/runs/baseline/checkpoint-1800/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/baseline/checkpoint-1800/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/baseline/checkpoint-1800/tokenizer_config.json b/runs/baseline/checkpoint-1800/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/baseline/checkpoint-1800/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/baseline/checkpoint-1800/trainer_state.json b/runs/baseline/checkpoint-1800/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..b4b06705b9f0abc5b4f2960828a410f5a9e4c9c7
--- /dev/null
+++ b/runs/baseline/checkpoint-1800/trainer_state.json
@@ -0,0 +1,12826 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 25.0,
+ "eval_steps": 500,
+ "global_step": 1800,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.12165174633264542,
+ "learning_rate": 0.0,
+ "loss": 12.027063369750977,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.12109383195638657,
+ "learning_rate": 3e-05,
+ "loss": 12.028635025024414,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.12690064311027527,
+ "learning_rate": 6e-05,
+ "loss": 11.894524574279785,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.1537286937236786,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.672262191772461,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.11778244376182556,
+ "learning_rate": 0.00012,
+ "loss": 11.433671951293945,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.10874935984611511,
+ "learning_rate": 0.00015000000000000001,
+ "loss": 11.270309448242188,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.10641921311616898,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 11.155481338500977,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.1048506498336792,
+ "learning_rate": 0.00021000000000000004,
+ "loss": 11.066675186157227,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.10546834021806717,
+ "learning_rate": 0.00024,
+ "loss": 10.952295303344727,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.10614268481731415,
+ "learning_rate": 0.00027,
+ "loss": 10.811386108398438,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10529010742902756,
+ "learning_rate": 0.00030000000000000003,
+ "loss": 10.657712936401367,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.1051444336771965,
+ "learning_rate": 0.00033,
+ "loss": 10.47591781616211,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10420846194028854,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 10.280141830444336,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10323625057935715,
+ "learning_rate": 0.00039000000000000005,
+ "loss": 10.076366424560547,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10255401581525803,
+ "learning_rate": 0.00042000000000000007,
+ "loss": 9.856939315795898,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10151027888059616,
+ "learning_rate": 0.00045,
+ "loss": 9.632134437561035,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.1006326898932457,
+ "learning_rate": 0.00048,
+ "loss": 9.401637077331543,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.09820562601089478,
+ "learning_rate": 0.00051,
+ "loss": 9.177400588989258,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.0956626906991005,
+ "learning_rate": 0.00054,
+ "loss": 8.95518684387207,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.0922091156244278,
+ "learning_rate": 0.00057,
+ "loss": 8.75066089630127,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.08593932539224625,
+ "learning_rate": 0.0006000000000000001,
+ "loss": 8.557976722717285,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.08014766871929169,
+ "learning_rate": 0.00063,
+ "loss": 8.360604286193848,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.07381244748830795,
+ "learning_rate": 0.00066,
+ "loss": 8.210375785827637,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.06455052644014359,
+ "learning_rate": 0.0006900000000000001,
+ "loss": 8.07168960571289,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.054115474224090576,
+ "learning_rate": 0.0007199999999999999,
+ "loss": 7.960069179534912,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.043575264513492584,
+ "learning_rate": 0.00075,
+ "loss": 7.850685119628906,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.02976679615676403,
+ "learning_rate": 0.0007800000000000001,
+ "loss": 7.824377059936523,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.020552165806293488,
+ "learning_rate": 0.0008100000000000001,
+ "loss": 7.7695417404174805,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.018530743196606636,
+ "learning_rate": 0.0008400000000000001,
+ "loss": 7.782247543334961,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.02716820128262043,
+ "learning_rate": 0.00087,
+ "loss": 7.754945278167725,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.03638409823179245,
+ "learning_rate": 0.0009,
+ "loss": 7.799381256103516,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.041579652577638626,
+ "learning_rate": 0.00093,
+ "loss": 7.776629447937012,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.0304960235953331,
+ "learning_rate": 0.00096,
+ "loss": 7.749439239501953,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.053728822618722916,
+ "learning_rate": 0.00099,
+ "loss": 7.745059490203857,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.036322493106126785,
+ "learning_rate": 0.00102,
+ "loss": 7.752074718475342,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.03232959285378456,
+ "learning_rate": 0.00105,
+ "loss": 7.704257011413574,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.03246184438467026,
+ "learning_rate": 0.00108,
+ "loss": 7.667045593261719,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.036655545234680176,
+ "learning_rate": 0.00111,
+ "loss": 7.623445987701416,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.021042725071310997,
+ "learning_rate": 0.00114,
+ "loss": 7.582821846008301,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.022982975468039513,
+ "learning_rate": 0.00117,
+ "loss": 7.58036994934082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.024781864136457443,
+ "learning_rate": 0.0012000000000000001,
+ "loss": 7.5903000831604,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.03667089343070984,
+ "learning_rate": 0.00123,
+ "loss": 7.588233947753906,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.04083017632365227,
+ "learning_rate": 0.00126,
+ "loss": 7.548343658447266,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.02585385926067829,
+ "learning_rate": 0.00129,
+ "loss": 7.604833602905273,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.018793931230902672,
+ "learning_rate": 0.00132,
+ "loss": 7.56705379486084,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.025538936257362366,
+ "learning_rate": 0.00135,
+ "loss": 7.562107086181641,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.02819318138062954,
+ "learning_rate": 0.0013800000000000002,
+ "loss": 7.517213821411133,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.016239255666732788,
+ "learning_rate": 0.00141,
+ "loss": 7.4780755043029785,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.024470701813697815,
+ "learning_rate": 0.0014399999999999999,
+ "loss": 7.505887985229492,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.015499847941100597,
+ "learning_rate": 0.00147,
+ "loss": 7.484469890594482,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.017785200849175453,
+ "learning_rate": 0.0015,
+ "loss": 7.481822967529297,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.015952805057168007,
+ "learning_rate": 0.0015300000000000001,
+ "loss": 7.484804153442383,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.015771260485053062,
+ "learning_rate": 0.0015600000000000002,
+ "loss": 7.4495439529418945,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.024370530620217323,
+ "learning_rate": 0.00159,
+ "loss": 7.471208572387695,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.018173737451434135,
+ "learning_rate": 0.0016200000000000001,
+ "loss": 7.438447952270508,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.020564094185829163,
+ "learning_rate": 0.0016500000000000002,
+ "loss": 7.437922954559326,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.016154944896697998,
+ "learning_rate": 0.0016800000000000003,
+ "loss": 7.429378032684326,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.016817007213830948,
+ "learning_rate": 0.00171,
+ "loss": 7.406428337097168,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.03460947796702385,
+ "learning_rate": 0.00174,
+ "loss": 7.393331527709961,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.15410488843917847,
+ "learning_rate": 0.0017699999999999999,
+ "loss": 7.645856857299805,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.07443148642778397,
+ "learning_rate": 0.0018,
+ "loss": 7.549508094787598,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.0507572740316391,
+ "learning_rate": 0.00183,
+ "loss": 7.517921447753906,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.09682746231555939,
+ "learning_rate": 0.00186,
+ "loss": 7.543978691101074,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.2571517527103424,
+ "learning_rate": 0.00189,
+ "loss": 7.768949508666992,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.0317784883081913,
+ "learning_rate": 0.00192,
+ "loss": 7.509063720703125,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.0353970006108284,
+ "learning_rate": 0.0019500000000000001,
+ "loss": 7.522322654724121,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.03663994371891022,
+ "learning_rate": 0.00198,
+ "loss": 7.504723072052002,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.025481393560767174,
+ "learning_rate": 0.00201,
+ "loss": 7.452597141265869,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.03099915198981762,
+ "learning_rate": 0.00204,
+ "loss": 7.455441474914551,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.022921698167920113,
+ "learning_rate": 0.00207,
+ "loss": 7.4454827308654785,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.032388731837272644,
+ "learning_rate": 0.0021,
+ "loss": 7.467103958129883,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.022428609430789948,
+ "learning_rate": 0.00213,
+ "loss": 7.467540264129639,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.469719409942627,
+ "eval_runtime": 50.9714,
+ "eval_samples_per_second": 47.909,
+ "eval_steps_per_second": 1.511,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.017480911687016487,
+ "learning_rate": 0.00216,
+ "loss": 7.455810546875,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.016675131395459175,
+ "learning_rate": 0.00219,
+ "loss": 7.474848747253418,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.017450228333473206,
+ "learning_rate": 0.00222,
+ "loss": 7.442611217498779,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.016823675483465195,
+ "learning_rate": 0.0022500000000000003,
+ "loss": 7.454628944396973,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.016717543825507164,
+ "learning_rate": 0.00228,
+ "loss": 7.425317764282227,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.01644429750740528,
+ "learning_rate": 0.00231,
+ "loss": 7.4102864265441895,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.013104400597512722,
+ "learning_rate": 0.00234,
+ "loss": 7.4559173583984375,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.01175684668123722,
+ "learning_rate": 0.00237,
+ "loss": 7.438898086547852,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.013736684806644917,
+ "learning_rate": 0.0024000000000000002,
+ "loss": 7.449005603790283,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01176756713539362,
+ "learning_rate": 0.0024300000000000003,
+ "loss": 7.428421974182129,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.011860918253660202,
+ "learning_rate": 0.00246,
+ "loss": 7.411491394042969,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.010093477554619312,
+ "learning_rate": 0.00249,
+ "loss": 7.39311408996582,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.009453288279473782,
+ "learning_rate": 0.00252,
+ "loss": 7.387656211853027,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.010311692021787167,
+ "learning_rate": 0.00255,
+ "loss": 7.401076316833496,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.010053195990622044,
+ "learning_rate": 0.00258,
+ "loss": 7.39655876159668,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.009758037514984608,
+ "learning_rate": 0.00261,
+ "loss": 7.423122882843018,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008394946344196796,
+ "learning_rate": 0.00264,
+ "loss": 7.383601188659668,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.008457187563180923,
+ "learning_rate": 0.00267,
+ "loss": 7.400934219360352,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.010601334273815155,
+ "learning_rate": 0.0027,
+ "loss": 7.400241851806641,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.009009373374283314,
+ "learning_rate": 0.0027300000000000002,
+ "loss": 7.377348899841309,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.00737508200109005,
+ "learning_rate": 0.0027600000000000003,
+ "loss": 7.385881423950195,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.0075711836107075214,
+ "learning_rate": 0.0027900000000000004,
+ "loss": 7.348570823669434,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.00846901722252369,
+ "learning_rate": 0.00282,
+ "loss": 7.364287853240967,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.007160635199397802,
+ "learning_rate": 0.00285,
+ "loss": 7.3881635665893555,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.00682067358866334,
+ "learning_rate": 0.0028799999999999997,
+ "loss": 7.348663330078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.009108101949095726,
+ "learning_rate": 0.00291,
+ "loss": 7.356302261352539,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.015828927978873253,
+ "learning_rate": 0.00294,
+ "loss": 7.347576141357422,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.009683096781373024,
+ "learning_rate": 0.00297,
+ "loss": 7.323967933654785,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.010833270847797394,
+ "learning_rate": 0.003,
+ "loss": 7.3139495849609375,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.014905400574207306,
+ "learning_rate": 0.002999999401781161,
+ "loss": 7.302030086517334,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.023674853146076202,
+ "learning_rate": 0.002999997607125127,
+ "loss": 7.311811923980713,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.050460800528526306,
+ "learning_rate": 0.002999994616033342,
+ "loss": 7.41972541809082,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.02193174883723259,
+ "learning_rate": 0.002999990428508218,
+ "loss": 7.294468879699707,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.036261629313230515,
+ "learning_rate": 0.0029999850445531283,
+ "loss": 7.295568466186523,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.02734929881989956,
+ "learning_rate": 0.002999978464172409,
+ "loss": 7.309239864349365,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.017884107306599617,
+ "learning_rate": 0.002999970687371363,
+ "loss": 7.2427592277526855,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01700308918952942,
+ "learning_rate": 0.0029999617141562573,
+ "loss": 7.241730213165283,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.0209935512393713,
+ "learning_rate": 0.0029999515445343194,
+ "loss": 7.262085437774658,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.023207105696201324,
+ "learning_rate": 0.0029999401785137437,
+ "loss": 7.25450325012207,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.01907353848218918,
+ "learning_rate": 0.002999927616103687,
+ "loss": 7.23149299621582,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.016149666160345078,
+ "learning_rate": 0.0029999138573142715,
+ "loss": 7.210232734680176,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.017513928934931755,
+ "learning_rate": 0.002999898902156582,
+ "loss": 7.218451976776123,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.015002547763288021,
+ "learning_rate": 0.002999882750642667,
+ "loss": 7.169316291809082,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.014233081601560116,
+ "learning_rate": 0.0029998654027855406,
+ "loss": 7.194825172424316,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.01283260714262724,
+ "learning_rate": 0.0029998468585991783,
+ "loss": 7.15669059753418,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.014416799880564213,
+ "learning_rate": 0.0029998271180985223,
+ "loss": 7.143369674682617,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.014337058179080486,
+ "learning_rate": 0.0029998061812994757,
+ "loss": 7.127734184265137,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.016406210139393806,
+ "learning_rate": 0.002999784048218908,
+ "loss": 7.086504936218262,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.01680421642959118,
+ "learning_rate": 0.002999760718874652,
+ "loss": 7.089420318603516,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.010665548034012318,
+ "learning_rate": 0.0029997361932855016,
+ "loss": 7.102463722229004,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013479137793183327,
+ "learning_rate": 0.002999710471471218,
+ "loss": 7.079098701477051,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.012408486567437649,
+ "learning_rate": 0.0029996835534525246,
+ "loss": 7.0790114402771,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.012035196647047997,
+ "learning_rate": 0.0029996554392511094,
+ "loss": 7.058756351470947,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.00899985246360302,
+ "learning_rate": 0.0029996261288896225,
+ "loss": 7.0739240646362305,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.01072906143963337,
+ "learning_rate": 0.002999595622391679,
+ "loss": 7.085537910461426,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.01047342736274004,
+ "learning_rate": 0.0029995639197818577,
+ "loss": 7.060049057006836,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.02342618815600872,
+ "learning_rate": 0.0029995310210857006,
+ "loss": 7.039482116699219,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.011256859637796879,
+ "learning_rate": 0.0029994969263297134,
+ "loss": 7.040412902832031,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.008407019078731537,
+ "learning_rate": 0.0029994616355413666,
+ "loss": 7.0364274978637695,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.011006148532032967,
+ "learning_rate": 0.0029994251487490916,
+ "loss": 6.991468906402588,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.009103255346417427,
+ "learning_rate": 0.0029993874659822875,
+ "loss": 7.001059532165527,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.007875788025557995,
+ "learning_rate": 0.0029993485872713125,
+ "loss": 6.9974260330200195,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.007872611284255981,
+ "learning_rate": 0.0029993085126474913,
+ "loss": 6.980216979980469,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.010071957483887672,
+ "learning_rate": 0.0029992672421431115,
+ "loss": 6.97799015045166,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.01209618803113699,
+ "learning_rate": 0.0029992247757914237,
+ "loss": 6.97202205657959,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.014884858392179012,
+ "learning_rate": 0.0029991811136266425,
+ "loss": 6.976570129394531,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.014953463338315487,
+ "learning_rate": 0.0029991362556839453,
+ "loss": 6.971050262451172,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.008371063508093357,
+ "learning_rate": 0.002999090201999474,
+ "loss": 6.955556869506836,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.011926881037652493,
+ "learning_rate": 0.002999042952610332,
+ "loss": 6.943947792053223,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.013422893360257149,
+ "learning_rate": 0.002998994507554589,
+ "loss": 6.910022735595703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.009672224521636963,
+ "learning_rate": 0.002998944866871275,
+ "loss": 6.8994855880737305,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.012590581551194191,
+ "learning_rate": 0.0029988940306003853,
+ "loss": 6.88709020614624,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.945146083831787,
+ "eval_runtime": 51.5801,
+ "eval_samples_per_second": 47.344,
+ "eval_steps_per_second": 1.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.016554836183786392,
+ "learning_rate": 0.002998841998782877,
+ "loss": 6.916609764099121,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.011085660196840763,
+ "learning_rate": 0.0029987887714606723,
+ "loss": 6.884910583496094,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.008596168830990791,
+ "learning_rate": 0.0029987343486766547,
+ "loss": 6.900066375732422,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.014487302862107754,
+ "learning_rate": 0.0029986787304746715,
+ "loss": 6.8693342208862305,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.015636997297406197,
+ "learning_rate": 0.002998621916899534,
+ "loss": 6.903670787811279,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.013275462202727795,
+ "learning_rate": 0.0029985639079970153,
+ "loss": 6.889361381530762,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.009044744074344635,
+ "learning_rate": 0.002998504703813852,
+ "loss": 6.8549981117248535,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.011327696964144707,
+ "learning_rate": 0.002998444304397745,
+ "loss": 6.856151103973389,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.01310331467539072,
+ "learning_rate": 0.0029983827097973554,
+ "loss": 6.834944725036621,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.010003753937780857,
+ "learning_rate": 0.00299831992006231,
+ "loss": 6.825613975524902,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.009928115643560886,
+ "learning_rate": 0.0029982559352431967,
+ "loss": 6.828935146331787,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.012302150949835777,
+ "learning_rate": 0.0029981907553915674,
+ "loss": 6.821207046508789,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.020555168390274048,
+ "learning_rate": 0.002998124380559936,
+ "loss": 6.8184895515441895,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.011417638510465622,
+ "learning_rate": 0.0029980568108017797,
+ "loss": 6.788754463195801,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.010772978886961937,
+ "learning_rate": 0.002997988046171538,
+ "loss": 6.7861433029174805,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.012753859162330627,
+ "learning_rate": 0.002997918086724614,
+ "loss": 6.788776397705078,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.011327114887535572,
+ "learning_rate": 0.0029978469325173717,
+ "loss": 6.779160022735596,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01235753484070301,
+ "learning_rate": 0.002997774583607139,
+ "loss": 6.799554824829102,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.01417156495153904,
+ "learning_rate": 0.0029977010400522074,
+ "loss": 6.77957820892334,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.021321553736925125,
+ "learning_rate": 0.002997626301911828,
+ "loss": 6.785632133483887,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.03614472970366478,
+ "learning_rate": 0.002997550369246217,
+ "loss": 6.836414337158203,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03934052586555481,
+ "learning_rate": 0.002997473242116551,
+ "loss": 6.854659557342529,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.016240060329437256,
+ "learning_rate": 0.0029973949205849707,
+ "loss": 6.789690971374512,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.031164905056357384,
+ "learning_rate": 0.0029973154047145782,
+ "loss": 6.8363776206970215,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.013051844201982021,
+ "learning_rate": 0.002997234694569438,
+ "loss": 6.803576469421387,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.022611448541283607,
+ "learning_rate": 0.002997152790214577,
+ "loss": 6.762107849121094,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.017812907695770264,
+ "learning_rate": 0.0029970696917159833,
+ "loss": 6.776218414306641,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.015453332103788853,
+ "learning_rate": 0.0029969853991406088,
+ "loss": 6.742618083953857,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.020630091428756714,
+ "learning_rate": 0.0029968999125563655,
+ "loss": 6.765475749969482,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.014672341756522655,
+ "learning_rate": 0.0029968132320321297,
+ "loss": 6.754009246826172,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.016540220007300377,
+ "learning_rate": 0.0029967253576377376,
+ "loss": 6.734752655029297,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.013495649211108685,
+ "learning_rate": 0.0029966362894439874,
+ "loss": 6.752676010131836,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015936989337205887,
+ "learning_rate": 0.0029965460275226405,
+ "loss": 6.725708961486816,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010431243106722832,
+ "learning_rate": 0.0029964545719464194,
+ "loss": 6.698376655578613,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01669321395456791,
+ "learning_rate": 0.0029963619227890083,
+ "loss": 6.711770534515381,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.013308997265994549,
+ "learning_rate": 0.0029962680801250514,
+ "loss": 6.705286979675293,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011648476123809814,
+ "learning_rate": 0.002996173044030159,
+ "loss": 6.690052509307861,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.03942161798477173,
+ "learning_rate": 0.0029960768145808965,
+ "loss": 6.711885452270508,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.01571274921298027,
+ "learning_rate": 0.002995979391854797,
+ "loss": 6.715847969055176,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.00829714722931385,
+ "learning_rate": 0.0029958807759303515,
+ "loss": 6.685308933258057,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.01374727487564087,
+ "learning_rate": 0.0029957809668870114,
+ "loss": 6.695744514465332,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.010857059620320797,
+ "learning_rate": 0.0029956799648051934,
+ "loss": 6.658294677734375,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.008897246792912483,
+ "learning_rate": 0.002995577769766272,
+ "loss": 6.642439842224121,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015880772843956947,
+ "learning_rate": 0.002995474381852584,
+ "loss": 6.674907684326172,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.021623898297548294,
+ "learning_rate": 0.0029953698011474265,
+ "loss": 6.66236686706543,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.03084627166390419,
+ "learning_rate": 0.0029952640277350596,
+ "loss": 6.666715621948242,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.040561504662036896,
+ "learning_rate": 0.002995157061700702,
+ "loss": 6.649062156677246,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.031126515939831734,
+ "learning_rate": 0.0029950489031305345,
+ "loss": 6.640542030334473,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.00872360821813345,
+ "learning_rate": 0.0029949395521116987,
+ "loss": 6.635221481323242,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.023432321846485138,
+ "learning_rate": 0.0029948290087322966,
+ "loss": 6.642196178436279,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.021096013486385345,
+ "learning_rate": 0.0029947172730813913,
+ "loss": 6.643655300140381,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.014494858682155609,
+ "learning_rate": 0.002994604345249006,
+ "loss": 6.612720966339111,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.05225599184632301,
+ "learning_rate": 0.0029944902253261248,
+ "loss": 6.60659646987915,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.014311269856989384,
+ "learning_rate": 0.0029943749134046915,
+ "loss": 6.54373025894165,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.010777472518384457,
+ "learning_rate": 0.0029942584095776115,
+ "loss": 6.599837303161621,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.012498079799115658,
+ "learning_rate": 0.002994140713938749,
+ "loss": 6.589657783508301,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.015775132924318314,
+ "learning_rate": 0.00299402182658293,
+ "loss": 6.577511787414551,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.017978230491280556,
+ "learning_rate": 0.0029939017476059396,
+ "loss": 6.58717155456543,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.026807745918631554,
+ "learning_rate": 0.0029937804771045236,
+ "loss": 6.578145503997803,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.04853282868862152,
+ "learning_rate": 0.0029936580151763872,
+ "loss": 6.609755516052246,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.1057916209101677,
+ "learning_rate": 0.002993534361920195,
+ "loss": 6.870367050170898,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.04925911873579025,
+ "learning_rate": 0.002993409517435574,
+ "loss": 6.871556282043457,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.05288556218147278,
+ "learning_rate": 0.0029932834818231077,
+ "loss": 6.808315753936768,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.03831710293889046,
+ "learning_rate": 0.0029931562551843405,
+ "loss": 6.827617645263672,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.03665117174386978,
+ "learning_rate": 0.002993027837621778,
+ "loss": 6.767284870147705,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.026216736063361168,
+ "learning_rate": 0.0029928982292388826,
+ "loss": 6.733659744262695,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.024796854704618454,
+ "learning_rate": 0.002992767430140078,
+ "loss": 6.759577751159668,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.02410992607474327,
+ "learning_rate": 0.002992635440430747,
+ "loss": 6.701825141906738,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.019924134016036987,
+ "learning_rate": 0.0029925022602172308,
+ "loss": 6.677121162414551,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.022516867145895958,
+ "learning_rate": 0.00299236788960683,
+ "loss": 6.668622970581055,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.025068556889891624,
+ "learning_rate": 0.002992232328707806,
+ "loss": 6.67718505859375,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.02574208751320839,
+ "learning_rate": 0.002992095577629376,
+ "loss": 6.679594039916992,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.657555103302002,
+ "eval_runtime": 51.0199,
+ "eval_samples_per_second": 47.864,
+ "eval_steps_per_second": 1.509,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.017298126593232155,
+ "learning_rate": 0.0029919576364817184,
+ "loss": 6.620774269104004,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.018840325996279716,
+ "learning_rate": 0.002991818505375971,
+ "loss": 6.613316535949707,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.020813113078475,
+ "learning_rate": 0.002991678184424228,
+ "loss": 6.621555328369141,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.026581479236483574,
+ "learning_rate": 0.0029915366737395437,
+ "loss": 6.613471031188965,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.02084030583500862,
+ "learning_rate": 0.0029913939734359306,
+ "loss": 6.559704780578613,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.01327502354979515,
+ "learning_rate": 0.00299125008362836,
+ "loss": 6.564910888671875,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.015091734007000923,
+ "learning_rate": 0.0029911050044327614,
+ "loss": 6.571262836456299,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.015463982708752155,
+ "learning_rate": 0.002990958735966023,
+ "loss": 6.565750598907471,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.013636826537549496,
+ "learning_rate": 0.0029908112783459894,
+ "loss": 6.57981014251709,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.010613273829221725,
+ "learning_rate": 0.0029906626316914655,
+ "loss": 6.538505554199219,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.01804237812757492,
+ "learning_rate": 0.0029905127961222127,
+ "loss": 6.5282511711120605,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.015131509862840176,
+ "learning_rate": 0.0029903617717589506,
+ "loss": 6.531335830688477,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.010518206283450127,
+ "learning_rate": 0.002990209558723358,
+ "loss": 6.521886825561523,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.011360022239387035,
+ "learning_rate": 0.002990056157138069,
+ "loss": 6.528301239013672,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.011556181125342846,
+ "learning_rate": 0.0029899015671266773,
+ "loss": 6.497917652130127,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.01365791168063879,
+ "learning_rate": 0.0029897457888137325,
+ "loss": 6.52849006652832,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.01091043371707201,
+ "learning_rate": 0.0029895888223247426,
+ "loss": 6.492197036743164,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.008484847843647003,
+ "learning_rate": 0.002989430667786173,
+ "loss": 6.497377395629883,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.00836879201233387,
+ "learning_rate": 0.0029892713253254466,
+ "loss": 6.495189666748047,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.009781497530639172,
+ "learning_rate": 0.0029891107950709406,
+ "loss": 6.454085826873779,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.009603648446500301,
+ "learning_rate": 0.0029889490771519934,
+ "loss": 6.499475479125977,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.007991855964064598,
+ "learning_rate": 0.0029887861716988975,
+ "loss": 6.460785865783691,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.008447187952697277,
+ "learning_rate": 0.0029886220788429025,
+ "loss": 6.504582405090332,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.010945498012006283,
+ "learning_rate": 0.0029884567987162156,
+ "loss": 6.49589729309082,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.011096215806901455,
+ "learning_rate": 0.0029882903314519997,
+ "loss": 6.460657119750977,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.013094739057123661,
+ "learning_rate": 0.0029881226771843747,
+ "loss": 6.462450981140137,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.015423037111759186,
+ "learning_rate": 0.0029879538360484165,
+ "loss": 6.45565128326416,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.0182222630828619,
+ "learning_rate": 0.002987783808180157,
+ "loss": 6.452227592468262,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.01981227844953537,
+ "learning_rate": 0.002987612593716585,
+ "loss": 6.477701663970947,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.026733700186014175,
+ "learning_rate": 0.002987440192795645,
+ "loss": 6.452212333679199,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.03504155948758125,
+ "learning_rate": 0.0029872666055562372,
+ "loss": 6.509665489196777,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.027970125898718834,
+ "learning_rate": 0.002987091832138217,
+ "loss": 6.487089157104492,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.013714591972529888,
+ "learning_rate": 0.0029869158726823974,
+ "loss": 6.469569206237793,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.016472695395350456,
+ "learning_rate": 0.0029867387273305453,
+ "loss": 6.461688995361328,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.018457358703017235,
+ "learning_rate": 0.0029865603962253833,
+ "loss": 6.490985870361328,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.011045100167393684,
+ "learning_rate": 0.0029863808795105896,
+ "loss": 6.403701305389404,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.014126897789537907,
+ "learning_rate": 0.0029862001773307972,
+ "loss": 6.379804611206055,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.014015335589647293,
+ "learning_rate": 0.0029860182898315954,
+ "loss": 6.404377460479736,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.010032658465206623,
+ "learning_rate": 0.002985835217159527,
+ "loss": 6.40291690826416,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.015469838865101337,
+ "learning_rate": 0.0029856509594620912,
+ "loss": 6.4414520263671875,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.011106910184025764,
+ "learning_rate": 0.0029854655168877397,
+ "loss": 6.396444320678711,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.01262320950627327,
+ "learning_rate": 0.0029852788895858816,
+ "loss": 6.385222434997559,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.012783976271748543,
+ "learning_rate": 0.002985091077706878,
+ "loss": 6.3818135261535645,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.01224067248404026,
+ "learning_rate": 0.0029849020814020468,
+ "loss": 6.37886905670166,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.010540086776018143,
+ "learning_rate": 0.0029847119008236576,
+ "loss": 6.37799072265625,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.012479913420975208,
+ "learning_rate": 0.0029845205361249366,
+ "loss": 6.365094184875488,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.010080077685415745,
+ "learning_rate": 0.0029843279874600622,
+ "loss": 6.371852874755859,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.013266298919916153,
+ "learning_rate": 0.002984134254984168,
+ "loss": 6.380039215087891,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.01014771405607462,
+ "learning_rate": 0.002983939338853339,
+ "loss": 6.370337963104248,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.008164579048752785,
+ "learning_rate": 0.0029837432392246185,
+ "loss": 6.331899642944336,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.01020442508161068,
+ "learning_rate": 0.002983545956255998,
+ "loss": 6.35077428817749,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.009898579679429531,
+ "learning_rate": 0.002983347490106426,
+ "loss": 6.371985912322998,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.009668919257819653,
+ "learning_rate": 0.002983147840935802,
+ "loss": 6.356776237487793,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.012560366652905941,
+ "learning_rate": 0.0029829470089049815,
+ "loss": 6.357911109924316,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.018435249105095863,
+ "learning_rate": 0.0029827449941757698,
+ "loss": 6.366821765899658,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.029046855866909027,
+ "learning_rate": 0.002982541796910927,
+ "loss": 6.367595672607422,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.043010637164115906,
+ "learning_rate": 0.002982337417274165,
+ "loss": 6.413588523864746,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.045470818877220154,
+ "learning_rate": 0.0029821318554301497,
+ "loss": 6.435015678405762,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.02148726023733616,
+ "learning_rate": 0.002981925111544498,
+ "loss": 6.330894470214844,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.026808688417077065,
+ "learning_rate": 0.00298171718578378,
+ "loss": 6.384839057922363,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.018495213240385056,
+ "learning_rate": 0.002981508078315517,
+ "loss": 6.36277961730957,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.025393597781658173,
+ "learning_rate": 0.0029812977893081834,
+ "loss": 6.3702239990234375,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020856289193034172,
+ "learning_rate": 0.002981086318931206,
+ "loss": 6.354547500610352,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.017827816307544708,
+ "learning_rate": 0.002980873667354962,
+ "loss": 6.339264869689941,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.02639714442193508,
+ "learning_rate": 0.002980659834750781,
+ "loss": 6.3254194259643555,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.015238096937537193,
+ "learning_rate": 0.002980444821290944,
+ "loss": 6.365842819213867,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.012863729149103165,
+ "learning_rate": 0.002980228627148684,
+ "loss": 6.356776714324951,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.014739899896085262,
+ "learning_rate": 0.0029800112524981845,
+ "loss": 6.332481384277344,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.013582171872258186,
+ "learning_rate": 0.0029797926975145797,
+ "loss": 6.311968803405762,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.012704637832939625,
+ "learning_rate": 0.0029795729623739555,
+ "loss": 6.321654319763184,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.012550560757517815,
+ "learning_rate": 0.0029793520472533495,
+ "loss": 6.298409461975098,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.011170165613293648,
+ "learning_rate": 0.002979129952330748,
+ "loss": 6.294830799102783,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.326757907867432,
+ "eval_runtime": 51.6566,
+ "eval_samples_per_second": 47.274,
+ "eval_steps_per_second": 1.491,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.010947446338832378,
+ "learning_rate": 0.0029789066777850897,
+ "loss": 6.286581039428711,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.010097995400428772,
+ "learning_rate": 0.002978682223796262,
+ "loss": 6.288301467895508,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.01224475260823965,
+ "learning_rate": 0.002978456590545104,
+ "loss": 6.264986991882324,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.009734537452459335,
+ "learning_rate": 0.0029782297782134045,
+ "loss": 6.304243087768555,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.01136956550180912,
+ "learning_rate": 0.0029780017869839013,
+ "loss": 6.244938373565674,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.00935293547809124,
+ "learning_rate": 0.002977772617040284,
+ "loss": 6.251852035522461,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.011118784546852112,
+ "learning_rate": 0.0029775422685671894,
+ "loss": 6.256633758544922,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.010837923735380173,
+ "learning_rate": 0.0029773107417502055,
+ "loss": 6.263411045074463,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.011920305900275707,
+ "learning_rate": 0.0029770780367758703,
+ "loss": 6.239311218261719,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.014117483049631119,
+ "learning_rate": 0.0029768441538316685,
+ "loss": 6.253314018249512,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.022502148523926735,
+ "learning_rate": 0.002976609093106036,
+ "loss": 6.246156692504883,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.03348720073699951,
+ "learning_rate": 0.002976372854788357,
+ "loss": 6.255914688110352,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.03427121415734291,
+ "learning_rate": 0.002976135439068965,
+ "loss": 6.289946556091309,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.02407355234026909,
+ "learning_rate": 0.002975896846139141,
+ "loss": 6.288534164428711,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.017861567437648773,
+ "learning_rate": 0.0029756570761911147,
+ "loss": 6.258049964904785,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.019980227574706078,
+ "learning_rate": 0.0029754161294180646,
+ "loss": 6.250895023345947,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.01443127728998661,
+ "learning_rate": 0.0029751740060141177,
+ "loss": 6.212200164794922,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.014975055120885372,
+ "learning_rate": 0.0029749307061743475,
+ "loss": 6.230568885803223,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01408296450972557,
+ "learning_rate": 0.0029746862300947773,
+ "loss": 6.243950366973877,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.011649327352643013,
+ "learning_rate": 0.0029744405779723766,
+ "loss": 6.197744846343994,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.013392188586294651,
+ "learning_rate": 0.002974193750005063,
+ "loss": 6.207277297973633,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.010902997106313705,
+ "learning_rate": 0.002973945746391701,
+ "loss": 6.166373252868652,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.012915392406284809,
+ "learning_rate": 0.0029736965673321032,
+ "loss": 6.186316013336182,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.010772830806672573,
+ "learning_rate": 0.0029734462130270273,
+ "loss": 6.172578811645508,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.010274911299347878,
+ "learning_rate": 0.002973194683678181,
+ "loss": 6.206851005554199,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.013210494071245193,
+ "learning_rate": 0.0029729419794882164,
+ "loss": 6.21115779876709,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.013372313231229782,
+ "learning_rate": 0.0029726881006607326,
+ "loss": 6.18442440032959,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.008105908520519733,
+ "learning_rate": 0.0029724330474002743,
+ "loss": 6.204354286193848,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.014167108573019505,
+ "learning_rate": 0.0029721768199123346,
+ "loss": 6.180662155151367,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.008339803665876389,
+ "learning_rate": 0.002971919418403351,
+ "loss": 6.150079727172852,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.015939386561512947,
+ "learning_rate": 0.0029716608430807073,
+ "loss": 6.175203323364258,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.010777109302580357,
+ "learning_rate": 0.0029714010941527327,
+ "loss": 6.166106224060059,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.011274735443294048,
+ "learning_rate": 0.002971140171828702,
+ "loss": 6.163111686706543,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013626808300614357,
+ "learning_rate": 0.002970878076318836,
+ "loss": 6.140850067138672,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.016907762736082077,
+ "learning_rate": 0.0029706148078343,
+ "loss": 6.192012786865234,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.018817441537976265,
+ "learning_rate": 0.002970350366587205,
+ "loss": 6.162238121032715,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.02829768694937229,
+ "learning_rate": 0.0029700847527906068,
+ "loss": 6.159066200256348,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.02034052088856697,
+ "learning_rate": 0.002969817966658505,
+ "loss": 6.164158344268799,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.01816369965672493,
+ "learning_rate": 0.0029695500084058447,
+ "loss": 6.132316589355469,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.013685261830687523,
+ "learning_rate": 0.0029692808782485146,
+ "loss": 6.152181625366211,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.009791149757802486,
+ "learning_rate": 0.0029690105764033485,
+ "loss": 6.138544082641602,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.014827069826424122,
+ "learning_rate": 0.002968739103088124,
+ "loss": 6.135766983032227,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.01391249243170023,
+ "learning_rate": 0.0029684664585215618,
+ "loss": 6.146172523498535,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.012729899026453495,
+ "learning_rate": 0.002968192642923326,
+ "loss": 6.122613430023193,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.014175458811223507,
+ "learning_rate": 0.002967917656514026,
+ "loss": 6.12574577331543,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.010958840139210224,
+ "learning_rate": 0.002967641499515213,
+ "loss": 6.111088752746582,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.015380816534161568,
+ "learning_rate": 0.0029673641721493817,
+ "loss": 6.100615501403809,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.025339096784591675,
+ "learning_rate": 0.00296708567463997,
+ "loss": 6.139659881591797,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.03695995733141899,
+ "learning_rate": 0.002966806007211358,
+ "loss": 6.160996437072754,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.0408046580851078,
+ "learning_rate": 0.0029665251700888686,
+ "loss": 6.173980712890625,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.025022024288773537,
+ "learning_rate": 0.0029662431634987674,
+ "loss": 6.145794868469238,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.027254154905676842,
+ "learning_rate": 0.002965959987668262,
+ "loss": 6.16928243637085,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.06191903352737427,
+ "learning_rate": 0.0029656756428255017,
+ "loss": 6.138525009155273,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0348956398665905,
+ "learning_rate": 0.002965390129199579,
+ "loss": 6.185477256774902,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.020859677344560623,
+ "learning_rate": 0.002965103447020526,
+ "loss": 6.148846626281738,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.022810814902186394,
+ "learning_rate": 0.0029648155965193174,
+ "loss": 6.1701178550720215,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.017981817945837975,
+ "learning_rate": 0.00296452657792787,
+ "loss": 6.1323771476745605,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.014905866235494614,
+ "learning_rate": 0.00296423639147904,
+ "loss": 6.123426914215088,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.01463468000292778,
+ "learning_rate": 0.002963945037406626,
+ "loss": 6.146122932434082,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.021616583690047264,
+ "learning_rate": 0.002963652515945366,
+ "loss": 6.093459606170654,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.01550968922674656,
+ "learning_rate": 0.00296335882733094,
+ "loss": 6.114929676055908,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.014597428031265736,
+ "learning_rate": 0.0029630639717999675,
+ "loss": 6.113945960998535,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.018249280750751495,
+ "learning_rate": 0.002962767949590008,
+ "loss": 6.130507469177246,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.014160148799419403,
+ "learning_rate": 0.0029624707609395614,
+ "loss": 6.086946487426758,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.015563832595944405,
+ "learning_rate": 0.002962172406088067,
+ "loss": 6.083332538604736,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.012240015901625156,
+ "learning_rate": 0.002961872885275904,
+ "loss": 6.0822906494140625,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.0141651201993227,
+ "learning_rate": 0.002961572198744391,
+ "loss": 6.049562454223633,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.010004645213484764,
+ "learning_rate": 0.0029612703467357866,
+ "loss": 6.067063808441162,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.011790835298597813,
+ "learning_rate": 0.0029609673294932856,
+ "loss": 6.062193870544434,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.011088060215115547,
+ "learning_rate": 0.002960663147261025,
+ "loss": 6.070950031280518,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.010387620888650417,
+ "learning_rate": 0.002960357800284078,
+ "loss": 6.032423973083496,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.012074458412826061,
+ "learning_rate": 0.002960051288808458,
+ "loss": 6.066788673400879,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.097111701965332,
+ "eval_runtime": 51.8742,
+ "eval_samples_per_second": 47.075,
+ "eval_steps_per_second": 1.484,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.013429952785372734,
+ "learning_rate": 0.0029597436130811153,
+ "loss": 6.022172451019287,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.01764705777168274,
+ "learning_rate": 0.0029594347733499387,
+ "loss": 6.029745101928711,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.02403099089860916,
+ "learning_rate": 0.002959124769863755,
+ "loss": 6.04707145690918,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.029030559584498405,
+ "learning_rate": 0.002958813602872328,
+ "loss": 6.05479097366333,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.029980644583702087,
+ "learning_rate": 0.0029585012726263595,
+ "loss": 6.068657875061035,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.021424276754260063,
+ "learning_rate": 0.0029581877793774885,
+ "loss": 6.062008857727051,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.018076395615935326,
+ "learning_rate": 0.002957873123378291,
+ "loss": 6.023279190063477,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.017688792198896408,
+ "learning_rate": 0.002957557304882279,
+ "loss": 6.014641284942627,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.02313626930117607,
+ "learning_rate": 0.0029572403241439026,
+ "loss": 6.026420593261719,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.04045062139630318,
+ "learning_rate": 0.0029569221814185467,
+ "loss": 6.007199764251709,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.015792209655046463,
+ "learning_rate": 0.0029566028769625334,
+ "loss": 6.017209053039551,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.01990479789674282,
+ "learning_rate": 0.002956282411033121,
+ "loss": 5.982841491699219,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.01191043108701706,
+ "learning_rate": 0.002955960783888503,
+ "loss": 6.0035929679870605,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.01732831820845604,
+ "learning_rate": 0.002955637995787807,
+ "loss": 6.025006294250488,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.032948266714811325,
+ "learning_rate": 0.0029553140469910995,
+ "loss": 6.007736682891846,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.011192727833986282,
+ "learning_rate": 0.00295498893775938,
+ "loss": 6.0022687911987305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014646840281784534,
+ "learning_rate": 0.0029546626683545824,
+ "loss": 5.976479530334473,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01589796133339405,
+ "learning_rate": 0.002954335239039576,
+ "loss": 5.980949878692627,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.015129734762012959,
+ "learning_rate": 0.0029540066500781657,
+ "loss": 6.0073394775390625,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.011140518821775913,
+ "learning_rate": 0.002953676901735088,
+ "loss": 5.976587295532227,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.0109480582177639,
+ "learning_rate": 0.0029533459942760174,
+ "loss": 6.001427173614502,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.013346790336072445,
+ "learning_rate": 0.0029530139279675583,
+ "loss": 5.96919059753418,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.012474581599235535,
+ "learning_rate": 0.002952680703077251,
+ "loss": 5.970426559448242,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01140566635876894,
+ "learning_rate": 0.002952346319873569,
+ "loss": 5.989171504974365,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.0137029392644763,
+ "learning_rate": 0.0029520107786259187,
+ "loss": 6.010115623474121,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015517761930823326,
+ "learning_rate": 0.002951674079604639,
+ "loss": 6.003649711608887,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.02173629403114319,
+ "learning_rate": 0.0029513362230810035,
+ "loss": 5.967074871063232,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02683914266526699,
+ "learning_rate": 0.002950997209327216,
+ "loss": 5.994434833526611,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.025865858420729637,
+ "learning_rate": 0.002950657038616415,
+ "loss": 6.010523796081543,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02544897235929966,
+ "learning_rate": 0.002950315711222668,
+ "loss": 5.994078159332275,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.028426019474864006,
+ "learning_rate": 0.0029499732274209777,
+ "loss": 6.020925521850586,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.026174750179052353,
+ "learning_rate": 0.0029496295874872768,
+ "loss": 5.985616683959961,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.02802138403058052,
+ "learning_rate": 0.00294928479169843,
+ "loss": 5.973659038543701,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.019348938018083572,
+ "learning_rate": 0.002948938840332233,
+ "loss": 5.990137577056885,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.016294041648507118,
+ "learning_rate": 0.0029485917336674115,
+ "loss": 5.962082862854004,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.016167785972356796,
+ "learning_rate": 0.002948243471983625,
+ "loss": 5.9594573974609375,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.014119232073426247,
+ "learning_rate": 0.00294789405556146,
+ "loss": 5.894467353820801,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.013805347494781017,
+ "learning_rate": 0.002947543484682436,
+ "loss": 5.964014530181885,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.014496497809886932,
+ "learning_rate": 0.002947191759629001,
+ "loss": 5.989920616149902,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.014813800342381,
+ "learning_rate": 0.0029468388806845333,
+ "loss": 5.9669413566589355,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.01153622381389141,
+ "learning_rate": 0.0029464848481333423,
+ "loss": 5.918018341064453,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.010525977239012718,
+ "learning_rate": 0.0029461296622606647,
+ "loss": 5.892477989196777,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.011410463601350784,
+ "learning_rate": 0.002945773323352667,
+ "loss": 5.925249099731445,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.010949424467980862,
+ "learning_rate": 0.0029454158316964464,
+ "loss": 5.90645694732666,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.00962804164737463,
+ "learning_rate": 0.002945057187580026,
+ "loss": 5.917054176330566,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.011517546139657497,
+ "learning_rate": 0.0029446973912923597,
+ "loss": 5.919445514678955,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01567666605114937,
+ "learning_rate": 0.0029443364431233285,
+ "loss": 5.884111404418945,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.016781140118837357,
+ "learning_rate": 0.002943974343363742,
+ "loss": 5.9252119064331055,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.02244546078145504,
+ "learning_rate": 0.0029436110923053375,
+ "loss": 5.933143615722656,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.028186792507767677,
+ "learning_rate": 0.00294324669024078,
+ "loss": 5.9079132080078125,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.028585223481059074,
+ "learning_rate": 0.002942881137463661,
+ "loss": 5.925262451171875,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.02781485952436924,
+ "learning_rate": 0.0029425144342685003,
+ "loss": 5.963315010070801,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.021486928686499596,
+ "learning_rate": 0.002942146580950744,
+ "loss": 5.930115699768066,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.016890818253159523,
+ "learning_rate": 0.002941777577806765,
+ "loss": 5.921439170837402,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.0169126745313406,
+ "learning_rate": 0.0029414074251338618,
+ "loss": 5.928220272064209,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.014659176580607891,
+ "learning_rate": 0.002941036123230261,
+ "loss": 5.934546947479248,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.015337883494794369,
+ "learning_rate": 0.002940663672395112,
+ "loss": 5.890145301818848,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01254988368600607,
+ "learning_rate": 0.002940290072928493,
+ "loss": 5.875492095947266,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.014658362604677677,
+ "learning_rate": 0.0029399153251314066,
+ "loss": 5.901272773742676,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.017342250794172287,
+ "learning_rate": 0.002939539429305779,
+ "loss": 5.875853538513184,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.016825608909130096,
+ "learning_rate": 0.0029391623857544643,
+ "loss": 5.941972732543945,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.016716884449124336,
+ "learning_rate": 0.002938784194781238,
+ "loss": 5.912367343902588,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.021707067266106606,
+ "learning_rate": 0.002938404856690803,
+ "loss": 5.895319938659668,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.030519135296344757,
+ "learning_rate": 0.002938024371788785,
+ "loss": 5.897165298461914,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.02572188340127468,
+ "learning_rate": 0.002937642740381733,
+ "loss": 5.879164695739746,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.016286717727780342,
+ "learning_rate": 0.002937259962777122,
+ "loss": 5.841421127319336,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.020656082779169083,
+ "learning_rate": 0.002936876039283347,
+ "loss": 5.9222259521484375,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.015350459143519402,
+ "learning_rate": 0.0029364909702097298,
+ "loss": 5.885847091674805,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01708570495247841,
+ "learning_rate": 0.0029361047558665136,
+ "loss": 5.872044563293457,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.013537376187741756,
+ "learning_rate": 0.0029357173965648634,
+ "loss": 5.866434097290039,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.013208637945353985,
+ "learning_rate": 0.002935328892616868,
+ "loss": 5.8624491691589355,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.012249554507434368,
+ "learning_rate": 0.002934939244335538,
+ "loss": 5.833232879638672,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.90985107421875,
+ "eval_runtime": 51.0696,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.01382362935692072,
+ "learning_rate": 0.0029345484520348062,
+ "loss": 5.835695266723633,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.012947693467140198,
+ "learning_rate": 0.0029341565160295262,
+ "loss": 5.826190948486328,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.013653872534632683,
+ "learning_rate": 0.002933763436635475,
+ "loss": 5.831789016723633,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.027596980333328247,
+ "learning_rate": 0.0029333692141693484,
+ "loss": 5.839963436126709,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01637990213930607,
+ "learning_rate": 0.0029329738489487648,
+ "loss": 5.805248260498047,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.02169182151556015,
+ "learning_rate": 0.0029325773412922627,
+ "loss": 5.838248252868652,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.03518201410770416,
+ "learning_rate": 0.0029321796915193015,
+ "loss": 5.834336280822754,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04983877018094063,
+ "learning_rate": 0.0029317808999502603,
+ "loss": 5.879856109619141,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05085664615035057,
+ "learning_rate": 0.0029313809669064377,
+ "loss": 5.946531295776367,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.03314858302474022,
+ "learning_rate": 0.0029309798927100532,
+ "loss": 5.909746170043945,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.034118980169296265,
+ "learning_rate": 0.0029305776776842453,
+ "loss": 5.903947830200195,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.03401247411966324,
+ "learning_rate": 0.0029301743221530705,
+ "loss": 5.936162948608398,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.043827857822179794,
+ "learning_rate": 0.0029297698264415047,
+ "loss": 5.953949451446533,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.06281330436468124,
+ "learning_rate": 0.002929364190875444,
+ "loss": 5.988225936889648,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.07216024398803711,
+ "learning_rate": 0.0029289574157817014,
+ "loss": 6.187741279602051,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.055205658078193665,
+ "learning_rate": 0.002928549501488008,
+ "loss": 6.177831649780273,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.04834286496043205,
+ "learning_rate": 0.0029281404483230133,
+ "loss": 6.209604263305664,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.0447777584195137,
+ "learning_rate": 0.0029277302566162837,
+ "loss": 6.146507263183594,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.05729430541396141,
+ "learning_rate": 0.0029273189266983035,
+ "loss": 6.280686378479004,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.0410807766020298,
+ "learning_rate": 0.002926906458900473,
+ "loss": 6.246824264526367,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.044607777148485184,
+ "learning_rate": 0.0029264928535551104,
+ "loss": 6.19956111907959,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.05141043663024902,
+ "learning_rate": 0.0029260781109954515,
+ "loss": 6.118578910827637,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.029248764738440514,
+ "learning_rate": 0.0029256622315556448,
+ "loss": 6.155360221862793,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.028403814882040024,
+ "learning_rate": 0.0029252452155707582,
+ "loss": 6.097352027893066,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.025872666388750076,
+ "learning_rate": 0.002924827063376773,
+ "loss": 6.072485446929932,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.018447555601596832,
+ "learning_rate": 0.0029244077753105874,
+ "loss": 6.077045440673828,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.023166058585047722,
+ "learning_rate": 0.0029239873517100147,
+ "loss": 6.040050983428955,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.019579261541366577,
+ "learning_rate": 0.002923565792913782,
+ "loss": 6.039424896240234,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.018369073048233986,
+ "learning_rate": 0.002923143099261531,
+ "loss": 6.021180152893066,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.014188461937010288,
+ "learning_rate": 0.0029227192710938195,
+ "loss": 5.9668049812316895,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.019567398354411125,
+ "learning_rate": 0.0029222943087521182,
+ "loss": 5.989189147949219,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.014190621674060822,
+ "learning_rate": 0.0029218682125788102,
+ "loss": 5.942726135253906,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01375526562333107,
+ "learning_rate": 0.002921440982917195,
+ "loss": 5.942777633666992,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.014203300699591637,
+ "learning_rate": 0.0029210126201114836,
+ "loss": 5.927928924560547,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.014231069944798946,
+ "learning_rate": 0.0029205831245067994,
+ "loss": 5.923510551452637,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.012771515175700188,
+ "learning_rate": 0.0029201524964491796,
+ "loss": 5.908803939819336,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.013399476185441017,
+ "learning_rate": 0.0029197207362855735,
+ "loss": 5.890079021453857,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.025862423703074455,
+ "learning_rate": 0.0029192878443638427,
+ "loss": 5.919907569885254,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.011315986514091492,
+ "learning_rate": 0.0029188538210327595,
+ "loss": 5.901529788970947,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.01235947571694851,
+ "learning_rate": 0.00291841866664201,
+ "loss": 5.903595924377441,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.012175127863883972,
+ "learning_rate": 0.0029179823815421893,
+ "loss": 5.868306636810303,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01195445004850626,
+ "learning_rate": 0.0029175449660848052,
+ "loss": 5.899898529052734,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.012801481410861015,
+ "learning_rate": 0.0029171064206222745,
+ "loss": 5.890621662139893,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.011909718625247478,
+ "learning_rate": 0.002916666745507926,
+ "loss": 5.86421012878418,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01162486057728529,
+ "learning_rate": 0.0029162259410959977,
+ "loss": 5.869139671325684,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.010923407040536404,
+ "learning_rate": 0.0029157840077416374,
+ "loss": 5.869789123535156,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.011943561024963856,
+ "learning_rate": 0.002915340945800904,
+ "loss": 5.873437881469727,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.010295520536601543,
+ "learning_rate": 0.002914896755630763,
+ "loss": 5.819953918457031,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.0093373479321599,
+ "learning_rate": 0.002914451437589092,
+ "loss": 5.829481601715088,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.01182442344725132,
+ "learning_rate": 0.002914004992034674,
+ "loss": 5.855249404907227,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.00939213763922453,
+ "learning_rate": 0.0029135574193272034,
+ "loss": 5.83876371383667,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.011648556217551231,
+ "learning_rate": 0.0029131087198272813,
+ "loss": 5.824906826019287,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.011227169074118137,
+ "learning_rate": 0.002912658893896416,
+ "loss": 5.847869873046875,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.011084851808845997,
+ "learning_rate": 0.002912207941897025,
+ "loss": 5.839183807373047,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.009923825040459633,
+ "learning_rate": 0.002911755864192432,
+ "loss": 5.818821907043457,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.011300318874418736,
+ "learning_rate": 0.002911302661146867,
+ "loss": 5.819517135620117,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.009915645234286785,
+ "learning_rate": 0.0029108483331254684,
+ "loss": 5.776939392089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.008965205401182175,
+ "learning_rate": 0.00291039288049428,
+ "loss": 5.824148178100586,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.011025927029550076,
+ "learning_rate": 0.0029099363036202515,
+ "loss": 5.832704544067383,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.010143411345779896,
+ "learning_rate": 0.0029094786028712387,
+ "loss": 5.802353382110596,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.010021256282925606,
+ "learning_rate": 0.002909019778616003,
+ "loss": 5.802136421203613,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.011356706731021404,
+ "learning_rate": 0.002908559831224211,
+ "loss": 5.807962417602539,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.010525298304855824,
+ "learning_rate": 0.002908098761066433,
+ "loss": 5.787661552429199,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.011147079057991505,
+ "learning_rate": 0.0029076365685141463,
+ "loss": 5.801015377044678,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.008777810260653496,
+ "learning_rate": 0.00290717325393973,
+ "loss": 5.779858589172363,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.011423609219491482,
+ "learning_rate": 0.002906708817716469,
+ "loss": 5.777166843414307,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.008639856241643429,
+ "learning_rate": 0.00290624326021855,
+ "loss": 5.747189521789551,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.009628898464143276,
+ "learning_rate": 0.002905776581821066,
+ "loss": 5.7666916847229,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.011988377198576927,
+ "learning_rate": 0.0029053087829000097,
+ "loss": 5.777951240539551,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.01693224348127842,
+ "learning_rate": 0.00290483986383228,
+ "loss": 5.818393707275391,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.020264796912670135,
+ "learning_rate": 0.002904369824995674,
+ "loss": 5.747611045837402,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.018283747136592865,
+ "learning_rate": 0.0029038986667688964,
+ "loss": 5.789663314819336,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.825676918029785,
+ "eval_runtime": 57.6908,
+ "eval_samples_per_second": 42.329,
+ "eval_steps_per_second": 1.335,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.011447477154433727,
+ "learning_rate": 0.002903426389531549,
+ "loss": 5.724975109100342,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.012539781630039215,
+ "learning_rate": 0.002902952993664138,
+ "loss": 5.747196674346924,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.012179837562143803,
+ "learning_rate": 0.002902478479548069,
+ "loss": 5.748995780944824,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.019896456971764565,
+ "learning_rate": 0.00290200284756565,
+ "loss": 5.751996994018555,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.022261742502450943,
+ "learning_rate": 0.002901526098100089,
+ "loss": 5.7516984939575195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02683536894619465,
+ "learning_rate": 0.0029010482315354946,
+ "loss": 5.716037750244141,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.018057072535157204,
+ "learning_rate": 0.0029005692482568747,
+ "loss": 5.778608322143555,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.011646891944110394,
+ "learning_rate": 0.002900089148650138,
+ "loss": 5.7308268547058105,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02230495773255825,
+ "learning_rate": 0.002899607933102092,
+ "loss": 5.768552780151367,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.01212464738637209,
+ "learning_rate": 0.002899125602000443,
+ "loss": 5.733015060424805,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.011764907278120518,
+ "learning_rate": 0.0028986421557337973,
+ "loss": 5.748104095458984,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.01593147963285446,
+ "learning_rate": 0.002898157594691658,
+ "loss": 5.701960563659668,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.016333773732185364,
+ "learning_rate": 0.0028976719192644277,
+ "loss": 5.716808795928955,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.017547868192195892,
+ "learning_rate": 0.002897185129843406,
+ "loss": 5.724377632141113,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.01662801019847393,
+ "learning_rate": 0.002896697226820791,
+ "loss": 5.737432956695557,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.015781691297888756,
+ "learning_rate": 0.002896208210589676,
+ "loss": 5.7376203536987305,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.014363214373588562,
+ "learning_rate": 0.0028957180815440534,
+ "loss": 5.7342610359191895,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.015048054046928883,
+ "learning_rate": 0.0028952268400788115,
+ "loss": 5.715073585510254,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.013753226026892662,
+ "learning_rate": 0.002894734486589734,
+ "loss": 5.73264217376709,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.015562563203275204,
+ "learning_rate": 0.002894241021473502,
+ "loss": 5.7072038650512695,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.01520819216966629,
+ "learning_rate": 0.0028937464451276916,
+ "loss": 5.683300018310547,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.013024724088609219,
+ "learning_rate": 0.002893250757950773,
+ "loss": 5.726273536682129,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.012760856188833714,
+ "learning_rate": 0.0028927539603421128,
+ "loss": 5.7249064445495605,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.013542579486966133,
+ "learning_rate": 0.0028922560527019723,
+ "loss": 5.70849084854126,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.014801754616200924,
+ "learning_rate": 0.002891757035431506,
+ "loss": 5.725185394287109,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.013000023551285267,
+ "learning_rate": 0.0028912569089327637,
+ "loss": 5.705799102783203,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.012919292785227299,
+ "learning_rate": 0.002890755673608688,
+ "loss": 5.674324989318848,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.015904201194643974,
+ "learning_rate": 0.002890253329863115,
+ "loss": 5.680788993835449,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.015467721968889236,
+ "learning_rate": 0.0028897498781007744,
+ "loss": 5.707531929016113,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.015157260000705719,
+ "learning_rate": 0.002889245318727288,
+ "loss": 5.689894676208496,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.014204306527972221,
+ "learning_rate": 0.00288873965214917,
+ "loss": 5.713001251220703,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.022045845165848732,
+ "learning_rate": 0.0028882328787738273,
+ "loss": 5.657631874084473,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.013380175456404686,
+ "learning_rate": 0.0028877249990095573,
+ "loss": 5.702192306518555,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.016698408871889114,
+ "learning_rate": 0.0028872160132655503,
+ "loss": 5.695206642150879,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.020001472905278206,
+ "learning_rate": 0.0028867059219518866,
+ "loss": 5.731742858886719,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.027350103482604027,
+ "learning_rate": 0.0028861947254795375,
+ "loss": 5.712314605712891,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.021862946450710297,
+ "learning_rate": 0.002885682424260365,
+ "loss": 5.697043418884277,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.019216977059841156,
+ "learning_rate": 0.0028851690187071204,
+ "loss": 5.698813438415527,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.02131095714867115,
+ "learning_rate": 0.002884654509233446,
+ "loss": 5.6918864250183105,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.016583887860178947,
+ "learning_rate": 0.0028841388962538724,
+ "loss": 5.711456298828125,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.016998684033751488,
+ "learning_rate": 0.0028836221801838194,
+ "loss": 5.675747871398926,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.01705755852162838,
+ "learning_rate": 0.0028831043614395966,
+ "loss": 5.697177410125732,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.014604352414608002,
+ "learning_rate": 0.0028825854404384,
+ "loss": 5.681950092315674,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.013561096973717213,
+ "learning_rate": 0.0028820654175983162,
+ "loss": 5.67357063293457,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.021149173378944397,
+ "learning_rate": 0.0028815442933383174,
+ "loss": 5.709477424621582,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.01431041955947876,
+ "learning_rate": 0.0028810220680782644,
+ "loss": 5.7012200355529785,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.01429709978401661,
+ "learning_rate": 0.0028804987422389045,
+ "loss": 5.689794540405273,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.032025713473558426,
+ "learning_rate": 0.0028799743162418712,
+ "loss": 5.6713762283325195,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.013168714940547943,
+ "learning_rate": 0.002879448790509686,
+ "loss": 5.685478210449219,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.014088666997849941,
+ "learning_rate": 0.0028789221654657556,
+ "loss": 5.645514965057373,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.013320053927600384,
+ "learning_rate": 0.002878394441534372,
+ "loss": 5.678678512573242,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.022180087864398956,
+ "learning_rate": 0.002877865619140712,
+ "loss": 5.660525321960449,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.01271889265626669,
+ "learning_rate": 0.002877335698710839,
+ "loss": 5.652164459228516,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.011617431417107582,
+ "learning_rate": 0.002876804680671701,
+ "loss": 5.6389946937561035,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.012040726840496063,
+ "learning_rate": 0.0028762725654511282,
+ "loss": 5.628744602203369,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.013148069381713867,
+ "learning_rate": 0.0028757393534778377,
+ "loss": 5.69610595703125,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.01826276071369648,
+ "learning_rate": 0.0028752050451814276,
+ "loss": 5.634238243103027,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.020393460988998413,
+ "learning_rate": 0.0028746696409923805,
+ "loss": 5.656968593597412,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.017072148621082306,
+ "learning_rate": 0.0028741331413420627,
+ "loss": 5.668219089508057,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.01694098301231861,
+ "learning_rate": 0.0028735955466627217,
+ "loss": 5.670073509216309,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.022760670632123947,
+ "learning_rate": 0.002873056857387488,
+ "loss": 5.649288177490234,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.03613442927598953,
+ "learning_rate": 0.0028725170739503734,
+ "loss": 5.723211288452148,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.04232329502701759,
+ "learning_rate": 0.002871976196786272,
+ "loss": 5.757920742034912,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.044210851192474365,
+ "learning_rate": 0.002871434226330959,
+ "loss": 5.702426910400391,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.044886402785778046,
+ "learning_rate": 0.0028708911630210905,
+ "loss": 5.742881774902344,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.022805195301771164,
+ "learning_rate": 0.002870347007294202,
+ "loss": 5.716353416442871,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02219395525753498,
+ "learning_rate": 0.00286980175958871,
+ "loss": 5.68503475189209,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.019240057095885277,
+ "learning_rate": 0.002869255420343911,
+ "loss": 5.700656890869141,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.01957295648753643,
+ "learning_rate": 0.0028687079899999807,
+ "loss": 5.70595645904541,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016326555982232094,
+ "learning_rate": 0.0028681594689979736,
+ "loss": 5.697229385375977,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015815885737538338,
+ "learning_rate": 0.0028676098577798234,
+ "loss": 5.7020792961120605,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.015865175053477287,
+ "learning_rate": 0.002867059156788342,
+ "loss": 5.648272514343262,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.7417449951171875,
+ "eval_runtime": 51.2263,
+ "eval_samples_per_second": 47.671,
+ "eval_steps_per_second": 1.503,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.016063183546066284,
+ "learning_rate": 0.0028665073664672185,
+ "loss": 5.65411376953125,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.014323881827294827,
+ "learning_rate": 0.0028659544872610214,
+ "loss": 5.629649639129639,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.014005238190293312,
+ "learning_rate": 0.002865400519615195,
+ "loss": 5.655135631561279,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.01385372132062912,
+ "learning_rate": 0.0028648454639760615,
+ "loss": 5.632213115692139,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.014073756523430347,
+ "learning_rate": 0.0028642893207908187,
+ "loss": 5.626431465148926,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.019551943987607956,
+ "learning_rate": 0.0028637320905075406,
+ "loss": 5.595447063446045,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.017571093514561653,
+ "learning_rate": 0.002863173773575179,
+ "loss": 5.661581039428711,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.01691303588449955,
+ "learning_rate": 0.0028626143704435585,
+ "loss": 5.641129493713379,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.01584242656826973,
+ "learning_rate": 0.0028620538815633807,
+ "loss": 5.640626907348633,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.014592235907912254,
+ "learning_rate": 0.0028614923073862213,
+ "loss": 5.611204147338867,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.016636405140161514,
+ "learning_rate": 0.0028609296483645304,
+ "loss": 5.614716053009033,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.01370833721011877,
+ "learning_rate": 0.002860365904951632,
+ "loss": 5.633011817932129,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.013218311592936516,
+ "learning_rate": 0.002859801077601725,
+ "loss": 5.588958740234375,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.016971616074442863,
+ "learning_rate": 0.0028592351667698786,
+ "loss": 5.635064125061035,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.015524905174970627,
+ "learning_rate": 0.0028586681729120393,
+ "loss": 5.611454963684082,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.014769157394766808,
+ "learning_rate": 0.0028581000964850215,
+ "loss": 5.62175178527832,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.014313343912363052,
+ "learning_rate": 0.002857530937946516,
+ "loss": 5.603628158569336,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.016748322173953056,
+ "learning_rate": 0.002856960697755083,
+ "loss": 5.617765426635742,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.013503300957381725,
+ "learning_rate": 0.0028563893763701538,
+ "loss": 5.636135578155518,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.012275655753910542,
+ "learning_rate": 0.002855816974252033,
+ "loss": 5.552533149719238,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.01288773212581873,
+ "learning_rate": 0.002855243491861894,
+ "loss": 5.620527267456055,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.013687008060514927,
+ "learning_rate": 0.002854668929661781,
+ "loss": 5.618595123291016,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.015678439289331436,
+ "learning_rate": 0.0028540932881146092,
+ "loss": 5.604031562805176,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.013567988760769367,
+ "learning_rate": 0.002853516567684162,
+ "loss": 5.549149513244629,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0157297495752573,
+ "learning_rate": 0.002852938768835092,
+ "loss": 5.618505477905273,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.011867019347846508,
+ "learning_rate": 0.0028523598920329233,
+ "loss": 5.579049110412598,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.01398563478142023,
+ "learning_rate": 0.0028517799377440443,
+ "loss": 5.597678184509277,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.015525273978710175,
+ "learning_rate": 0.0028511989064357148,
+ "loss": 5.601663589477539,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.01588466763496399,
+ "learning_rate": 0.002850616798576061,
+ "loss": 5.5650434494018555,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.014276969246566296,
+ "learning_rate": 0.0028500336146340772,
+ "loss": 5.607125282287598,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.013253689743578434,
+ "learning_rate": 0.002849449355079623,
+ "loss": 5.6299824714660645,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.01176661066710949,
+ "learning_rate": 0.0028488640203834263,
+ "loss": 5.560839653015137,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.011712565086781979,
+ "learning_rate": 0.0028482776110170813,
+ "loss": 5.618312835693359,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.013515647500753403,
+ "learning_rate": 0.002847690127453047,
+ "loss": 5.566764831542969,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.020177524536848068,
+ "learning_rate": 0.0028471015701646485,
+ "loss": 5.554571628570557,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.029197178781032562,
+ "learning_rate": 0.0028465119396260757,
+ "loss": 5.6097259521484375,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.027947140857577324,
+ "learning_rate": 0.0028459212363123824,
+ "loss": 5.65598201751709,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.021089809015393257,
+ "learning_rate": 0.0028453294606994887,
+ "loss": 5.67195987701416,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.019936271011829376,
+ "learning_rate": 0.0028447366132641773,
+ "loss": 5.609482765197754,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.0176444910466671,
+ "learning_rate": 0.002844142694484094,
+ "loss": 5.60432767868042,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.023154428228735924,
+ "learning_rate": 0.0028435477048377495,
+ "loss": 5.616270065307617,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.018391480669379234,
+ "learning_rate": 0.002842951644804515,
+ "loss": 5.600588321685791,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.013716557994484901,
+ "learning_rate": 0.002842354514864627,
+ "loss": 5.595095634460449,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.014372860081493855,
+ "learning_rate": 0.002841756315499181,
+ "loss": 5.584517478942871,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.014930845238268375,
+ "learning_rate": 0.002841157047190135,
+ "loss": 5.587005138397217,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.012631502002477646,
+ "learning_rate": 0.0028405567104203095,
+ "loss": 5.594094276428223,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.014677002094686031,
+ "learning_rate": 0.0028399553056733854,
+ "loss": 5.569582939147949,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.015558961778879166,
+ "learning_rate": 0.0028393528334339025,
+ "loss": 5.596535682678223,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.014322095550596714,
+ "learning_rate": 0.0028387492941872624,
+ "loss": 5.570324420928955,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.012913414277136326,
+ "learning_rate": 0.002838144688419726,
+ "loss": 5.573002338409424,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.01251310296356678,
+ "learning_rate": 0.0028375390166184137,
+ "loss": 5.550327777862549,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.013847017660737038,
+ "learning_rate": 0.0028369322792713026,
+ "loss": 5.54567813873291,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.013789358548820019,
+ "learning_rate": 0.0028363244768672313,
+ "loss": 5.571495056152344,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.01236631814390421,
+ "learning_rate": 0.0028357156098958954,
+ "loss": 5.569730758666992,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.013210169039666653,
+ "learning_rate": 0.002835105678847847,
+ "loss": 5.53984260559082,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.01604137383401394,
+ "learning_rate": 0.0028344946842144975,
+ "loss": 5.530598163604736,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.01708815060555935,
+ "learning_rate": 0.002833882626488114,
+ "loss": 5.529256820678711,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.017222166061401367,
+ "learning_rate": 0.0028332695061618197,
+ "loss": 5.564949035644531,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.01803218014538288,
+ "learning_rate": 0.002832655323729595,
+ "loss": 5.560049057006836,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.01658225618302822,
+ "learning_rate": 0.002832040079686276,
+ "loss": 5.553183555603027,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.016834143549203873,
+ "learning_rate": 0.002831423774527554,
+ "loss": 5.573678970336914,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.01856408454477787,
+ "learning_rate": 0.002830806408749974,
+ "loss": 5.539497375488281,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.026909375563263893,
+ "learning_rate": 0.002830187982850936,
+ "loss": 5.551433563232422,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.016046348959207535,
+ "learning_rate": 0.002829568497328696,
+ "loss": 5.566728591918945,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.018582617864012718,
+ "learning_rate": 0.0028289479526823614,
+ "loss": 5.543226718902588,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.0211537666618824,
+ "learning_rate": 0.0028283263494118943,
+ "loss": 5.560711860656738,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.013894101604819298,
+ "learning_rate": 0.0028277036880181095,
+ "loss": 5.528890132904053,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.017581885680556297,
+ "learning_rate": 0.0028270799690026725,
+ "loss": 5.57473087310791,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.01622696965932846,
+ "learning_rate": 0.002826455192868104,
+ "loss": 5.578961372375488,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.01966337114572525,
+ "learning_rate": 0.0028258293601177743,
+ "loss": 5.513942718505859,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021581366658210754,
+ "learning_rate": 0.002825202471255906,
+ "loss": 5.573669910430908,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.020468758419156075,
+ "learning_rate": 0.002824574526787571,
+ "loss": 5.516103744506836,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.626922130584717,
+ "eval_runtime": 54.2404,
+ "eval_samples_per_second": 45.022,
+ "eval_steps_per_second": 1.42,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.014597049914300442,
+ "learning_rate": 0.0028239455272186936,
+ "loss": 5.506453514099121,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.020277054980397224,
+ "learning_rate": 0.002823315473056048,
+ "loss": 5.535350799560547,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.019103925675153732,
+ "learning_rate": 0.0028226843648072562,
+ "loss": 5.522078037261963,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.07158710062503815,
+ "learning_rate": 0.0028220522029807913,
+ "loss": 5.531142234802246,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01958007924258709,
+ "learning_rate": 0.0028214189880859746,
+ "loss": 5.484055042266846,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.0185184795409441,
+ "learning_rate": 0.002820784720632976,
+ "loss": 5.511264801025391,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01976991444826126,
+ "learning_rate": 0.0028201494011328133,
+ "loss": 5.51821231842041,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.02188604325056076,
+ "learning_rate": 0.002819513030097352,
+ "loss": 5.483095169067383,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.019405223429203033,
+ "learning_rate": 0.002818875608039305,
+ "loss": 5.539701461791992,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.016907615587115288,
+ "learning_rate": 0.0028182371354722306,
+ "loss": 5.518251419067383,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.02160538360476494,
+ "learning_rate": 0.0028175976129105364,
+ "loss": 5.509786605834961,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.023125721141695976,
+ "learning_rate": 0.002816957040869473,
+ "loss": 5.535561561584473,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.022233380004763603,
+ "learning_rate": 0.0028163154198651372,
+ "loss": 5.553831100463867,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.017227552831172943,
+ "learning_rate": 0.002815672750414473,
+ "loss": 5.547080993652344,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.017122473567724228,
+ "learning_rate": 0.0028150290330352665,
+ "loss": 5.493778228759766,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.020641913637518883,
+ "learning_rate": 0.0028143842682461498,
+ "loss": 5.536645889282227,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.017383966594934464,
+ "learning_rate": 0.0028137384565665967,
+ "loss": 5.506233215332031,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.017969606444239616,
+ "learning_rate": 0.002813091598516928,
+ "loss": 5.524741172790527,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.01936613954603672,
+ "learning_rate": 0.0028124436946183047,
+ "loss": 5.460989952087402,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.017492495477199554,
+ "learning_rate": 0.0028117947453927313,
+ "loss": 5.487619876861572,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.018786396831274033,
+ "learning_rate": 0.0028111447513630545,
+ "loss": 5.516829013824463,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.020149117335677147,
+ "learning_rate": 0.0028104937130529624,
+ "loss": 5.460541725158691,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.024668103083968163,
+ "learning_rate": 0.0028098416309869857,
+ "loss": 5.497687339782715,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.015888608992099762,
+ "learning_rate": 0.002809188505690494,
+ "loss": 5.494459629058838,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.019327623769640923,
+ "learning_rate": 0.0028085343376896997,
+ "loss": 5.486351490020752,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.026855193078517914,
+ "learning_rate": 0.0028078791275116532,
+ "loss": 5.509825706481934,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.026340370997786522,
+ "learning_rate": 0.002807222875684246,
+ "loss": 5.527161598205566,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.02200835384428501,
+ "learning_rate": 0.002806565582736208,
+ "loss": 5.507146835327148,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.020420534536242485,
+ "learning_rate": 0.0028059072491971085,
+ "loss": 5.5162506103515625,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02493203990161419,
+ "learning_rate": 0.0028052478755973545,
+ "loss": 5.5252556800842285,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.021293973550200462,
+ "learning_rate": 0.0028045874624681925,
+ "loss": 5.523589134216309,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.016593236476182938,
+ "learning_rate": 0.0028039260103417045,
+ "loss": 5.500586032867432,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.018857723101973534,
+ "learning_rate": 0.002803263519750811,
+ "loss": 5.52933406829834,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.015930436551570892,
+ "learning_rate": 0.0028025999912292677,
+ "loss": 5.48813533782959,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.019427936524152756,
+ "learning_rate": 0.002801935425311669,
+ "loss": 5.46573543548584,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.018123053014278412,
+ "learning_rate": 0.0028012698225334427,
+ "loss": 5.499445915222168,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.02035410702228546,
+ "learning_rate": 0.002800603183430853,
+ "loss": 5.513540267944336,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.025838350877165794,
+ "learning_rate": 0.0027999355085409996,
+ "loss": 5.565637588500977,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.029621612280607224,
+ "learning_rate": 0.002799266798401816,
+ "loss": 5.5619215965271,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.028948379680514336,
+ "learning_rate": 0.0027985970535520685,
+ "loss": 5.551331520080566,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.022135930135846138,
+ "learning_rate": 0.0027979262745313604,
+ "loss": 5.527722358703613,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.028884069994091988,
+ "learning_rate": 0.0027972544618801256,
+ "loss": 5.554599761962891,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.02499174140393734,
+ "learning_rate": 0.00279658161613963,
+ "loss": 5.543351173400879,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.021131042391061783,
+ "learning_rate": 0.0027959077378519755,
+ "loss": 5.496039390563965,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.02957489714026451,
+ "learning_rate": 0.002795232827560092,
+ "loss": 5.552300453186035,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.022832119837403297,
+ "learning_rate": 0.0027945568858077433,
+ "loss": 5.488651275634766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.02074701525270939,
+ "learning_rate": 0.002793879913139523,
+ "loss": 5.491244792938232,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.019919803366065025,
+ "learning_rate": 0.002793201910100856,
+ "loss": 5.483855247497559,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.01858169585466385,
+ "learning_rate": 0.002792522877237998,
+ "loss": 5.50787353515625,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02085857465863228,
+ "learning_rate": 0.002791842815098032,
+ "loss": 5.512872695922852,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.01407450158149004,
+ "learning_rate": 0.002791161724228872,
+ "loss": 5.472590446472168,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.01672944612801075,
+ "learning_rate": 0.002790479605179262,
+ "loss": 5.469183921813965,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.016565700992941856,
+ "learning_rate": 0.0027897964584987713,
+ "loss": 5.493677139282227,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.017017360776662827,
+ "learning_rate": 0.0027891122847377996,
+ "loss": 5.464931488037109,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.019842486828565598,
+ "learning_rate": 0.002788427084447574,
+ "loss": 5.461293697357178,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.016210904344916344,
+ "learning_rate": 0.0027877408581801467,
+ "loss": 5.454716682434082,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.02064039371907711,
+ "learning_rate": 0.002787053606488398,
+ "loss": 5.466576099395752,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.018198711797595024,
+ "learning_rate": 0.0027863653299260356,
+ "loss": 5.447747230529785,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.015937626361846924,
+ "learning_rate": 0.00278567602904759,
+ "loss": 5.5055036544799805,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.01687259040772915,
+ "learning_rate": 0.002784985704408419,
+ "loss": 5.422250747680664,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.02622046135365963,
+ "learning_rate": 0.002784294356564705,
+ "loss": 5.466894626617432,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.018672186881303787,
+ "learning_rate": 0.0027836019860734535,
+ "loss": 5.48300838470459,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.02166241593658924,
+ "learning_rate": 0.0027829085934924958,
+ "loss": 5.477658271789551,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018426023423671722,
+ "learning_rate": 0.0027822141793804856,
+ "loss": 5.487186431884766,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.022229891270399094,
+ "learning_rate": 0.0027815187442968995,
+ "loss": 5.488579273223877,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.026930972933769226,
+ "learning_rate": 0.002780822288802037,
+ "loss": 5.492000102996826,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.02733265422284603,
+ "learning_rate": 0.00278012481345702,
+ "loss": 5.4663801193237305,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.02993253618478775,
+ "learning_rate": 0.0027794263188237922,
+ "loss": 5.543368816375732,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03152153640985489,
+ "learning_rate": 0.002778726805465118,
+ "loss": 5.565644264221191,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.031395044177770615,
+ "learning_rate": 0.002778026273944581,
+ "loss": 5.553359508514404,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.026603225618600845,
+ "learning_rate": 0.002777324724826589,
+ "loss": 5.552347183227539,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.024152139201760292,
+ "learning_rate": 0.002776622158676366,
+ "loss": 5.493257522583008,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.602636814117432,
+ "eval_runtime": 54.6199,
+ "eval_samples_per_second": 44.709,
+ "eval_steps_per_second": 1.41,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.02089761756360531,
+ "learning_rate": 0.0027759185760599576,
+ "loss": 5.461747646331787,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.018181778490543365,
+ "learning_rate": 0.0027752139775442278,
+ "loss": 5.492741584777832,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.020166238769888878,
+ "learning_rate": 0.0027745083636968575,
+ "loss": 5.476003646850586,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.015491694211959839,
+ "learning_rate": 0.0027738017350863488,
+ "loss": 5.425764560699463,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.022291697561740875,
+ "learning_rate": 0.0027730940922820182,
+ "loss": 5.45866584777832,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.018074683845043182,
+ "learning_rate": 0.002772385435854002,
+ "loss": 5.465143203735352,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.016872268170118332,
+ "learning_rate": 0.002771675766373251,
+ "loss": 5.446003437042236,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.023103339597582817,
+ "learning_rate": 0.0027709650844115336,
+ "loss": 5.411036491394043,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.027957158163189888,
+ "learning_rate": 0.0027702533905414324,
+ "loss": 5.452442169189453,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.019935399293899536,
+ "learning_rate": 0.002769540685336348,
+ "loss": 5.412145614624023,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.014351371675729752,
+ "learning_rate": 0.0027688269693704926,
+ "loss": 5.434524059295654,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.0152254244312644,
+ "learning_rate": 0.0027681122432188948,
+ "loss": 5.428544998168945,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.015071330592036247,
+ "learning_rate": 0.0027673965074573965,
+ "loss": 5.459929466247559,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.023415958508849144,
+ "learning_rate": 0.0027666797626626537,
+ "loss": 5.447159767150879,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.017228802666068077,
+ "learning_rate": 0.0027659620094121337,
+ "loss": 5.442273139953613,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.014184504747390747,
+ "learning_rate": 0.002765243248284118,
+ "loss": 5.448305130004883,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.024384526535868645,
+ "learning_rate": 0.0027645234798576988,
+ "loss": 5.4076337814331055,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.043152231723070145,
+ "learning_rate": 0.0027638027047127813,
+ "loss": 5.45344352722168,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.02889758162200451,
+ "learning_rate": 0.002763080923430081,
+ "loss": 5.489194393157959,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.02686968818306923,
+ "learning_rate": 0.002762358136591123,
+ "loss": 5.493659496307373,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.020467868074774742,
+ "learning_rate": 0.002761634344778245,
+ "loss": 5.410489559173584,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.019759224727749825,
+ "learning_rate": 0.002760909548574592,
+ "loss": 5.439385890960693,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.0233843382447958,
+ "learning_rate": 0.0027601837485641194,
+ "loss": 5.445036888122559,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.031753458082675934,
+ "learning_rate": 0.0027594569453315916,
+ "loss": 5.503895282745361,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.02089349366724491,
+ "learning_rate": 0.002758729139462581,
+ "loss": 5.45689582824707,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.03339357674121857,
+ "learning_rate": 0.002758000331543466,
+ "loss": 5.522361755371094,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019328996539115906,
+ "learning_rate": 0.002757270522161436,
+ "loss": 5.441373825073242,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019863856956362724,
+ "learning_rate": 0.002756539711904484,
+ "loss": 5.4633283615112305,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.02141754701733589,
+ "learning_rate": 0.0027558079013614116,
+ "loss": 5.453215599060059,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016591915860772133,
+ "learning_rate": 0.002755075091121824,
+ "loss": 5.463696479797363,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.01955028623342514,
+ "learning_rate": 0.0027543412817761345,
+ "loss": 5.411355018615723,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.016791662201285362,
+ "learning_rate": 0.0027536064739155596,
+ "loss": 5.459614276885986,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.019661981612443924,
+ "learning_rate": 0.00275287066813212,
+ "loss": 5.440699577331543,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.01927965320646763,
+ "learning_rate": 0.002752133865018642,
+ "loss": 5.409865856170654,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.016483580693602562,
+ "learning_rate": 0.002751396065168754,
+ "loss": 5.390999794006348,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016264773905277252,
+ "learning_rate": 0.0027506572691768888,
+ "loss": 5.418543815612793,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.01958460919559002,
+ "learning_rate": 0.00274991747763828,
+ "loss": 5.403289794921875,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.01861695572733879,
+ "learning_rate": 0.002749176691148965,
+ "loss": 5.431911468505859,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.029594263061881065,
+ "learning_rate": 0.002748434910305782,
+ "loss": 5.399242401123047,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.028574613854289055,
+ "learning_rate": 0.0027476921357063697,
+ "loss": 5.440446853637695,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.028612585738301277,
+ "learning_rate": 0.0027469483679491693,
+ "loss": 5.410346031188965,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.020704898983240128,
+ "learning_rate": 0.00274620360763342,
+ "loss": 5.44694709777832,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.020542364567518234,
+ "learning_rate": 0.002745457855359163,
+ "loss": 5.402653694152832,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.015762662515044212,
+ "learning_rate": 0.0027447111117272363,
+ "loss": 5.396594047546387,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022096984088420868,
+ "learning_rate": 0.002743963377339278,
+ "loss": 5.398744106292725,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.018791764974594116,
+ "learning_rate": 0.002743214652797724,
+ "loss": 5.397758483886719,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01642526686191559,
+ "learning_rate": 0.0027424649387058093,
+ "loss": 5.411459922790527,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.017931703478097916,
+ "learning_rate": 0.0027417142356675637,
+ "loss": 5.390264511108398,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018819602206349373,
+ "learning_rate": 0.0027409625442878153,
+ "loss": 5.407539367675781,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.019389411434531212,
+ "learning_rate": 0.0027402098651721886,
+ "loss": 5.386474609375,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.013645983301103115,
+ "learning_rate": 0.002739456198927104,
+ "loss": 5.373845100402832,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.013846895657479763,
+ "learning_rate": 0.002738701546159776,
+ "loss": 5.387155055999756,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013553125783801079,
+ "learning_rate": 0.002737945907478215,
+ "loss": 5.377801895141602,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.012771612964570522,
+ "learning_rate": 0.002737189283491225,
+ "loss": 5.344066143035889,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.020465301349759102,
+ "learning_rate": 0.002736431674808405,
+ "loss": 5.382771968841553,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.014798106625676155,
+ "learning_rate": 0.0027356730820401465,
+ "loss": 5.370522499084473,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.01566634140908718,
+ "learning_rate": 0.0027349135057976332,
+ "loss": 5.351980209350586,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.017485102638602257,
+ "learning_rate": 0.002734152946692843,
+ "loss": 5.355443477630615,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.016646096482872963,
+ "learning_rate": 0.0027333914053385444,
+ "loss": 5.34829044342041,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.017293352633714676,
+ "learning_rate": 0.0027326288823482964,
+ "loss": 5.3519487380981445,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.018873002380132675,
+ "learning_rate": 0.002731865378336452,
+ "loss": 5.348442077636719,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.015504784882068634,
+ "learning_rate": 0.002731100893918151,
+ "loss": 5.341490268707275,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.014270870015025139,
+ "learning_rate": 0.0027303354297093254,
+ "loss": 5.359251022338867,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.018820680677890778,
+ "learning_rate": 0.002729568986326695,
+ "loss": 5.390280723571777,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.021008187904953957,
+ "learning_rate": 0.0027288015643877708,
+ "loss": 5.338380336761475,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.016266655176877975,
+ "learning_rate": 0.0027280331645108498,
+ "loss": 5.347043037414551,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.01831045374274254,
+ "learning_rate": 0.0027272637873150185,
+ "loss": 5.326537132263184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02073794975876808,
+ "learning_rate": 0.002726493433420149,
+ "loss": 5.3607025146484375,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.022534754127264023,
+ "learning_rate": 0.0027257221034469033,
+ "loss": 5.353292465209961,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.021627362817525864,
+ "learning_rate": 0.002724949798016727,
+ "loss": 5.373323917388916,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.018226243555545807,
+ "learning_rate": 0.0027241765177518523,
+ "loss": 5.344578266143799,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.02032875455915928,
+ "learning_rate": 0.002723402263275298,
+ "loss": 5.329159259796143,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.44969367980957,
+ "eval_runtime": 51.2294,
+ "eval_samples_per_second": 47.668,
+ "eval_steps_per_second": 1.503,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.01613701693713665,
+ "learning_rate": 0.0027226270352108664,
+ "loss": 5.3024678230285645,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01940973661839962,
+ "learning_rate": 0.0027218508341831445,
+ "loss": 5.270181179046631,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.015840493142604828,
+ "learning_rate": 0.002721073660817504,
+ "loss": 5.261706352233887,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.017390675842761993,
+ "learning_rate": 0.0027202955157401,
+ "loss": 5.262207984924316,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.015184012241661549,
+ "learning_rate": 0.0027195163995778686,
+ "loss": 5.297155380249023,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.013417073525488377,
+ "learning_rate": 0.00271873631295853,
+ "loss": 5.279838562011719,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.013024317100644112,
+ "learning_rate": 0.0027179552565105864,
+ "loss": 5.265446186065674,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.01426338218152523,
+ "learning_rate": 0.002717173230863321,
+ "loss": 5.284180164337158,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.019604040309786797,
+ "learning_rate": 0.002716390236646797,
+ "loss": 5.286281108856201,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.019039476290345192,
+ "learning_rate": 0.0027156062744918587,
+ "loss": 5.290433406829834,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.019284334033727646,
+ "learning_rate": 0.0027148213450301313,
+ "loss": 5.262433052062988,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.0145154669880867,
+ "learning_rate": 0.0027140354488940175,
+ "loss": 5.24387264251709,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.014776417054235935,
+ "learning_rate": 0.0027132485867166992,
+ "loss": 5.282435417175293,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.012711729854345322,
+ "learning_rate": 0.0027124607591321377,
+ "loss": 5.246735572814941,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.013394835405051708,
+ "learning_rate": 0.0027116719667750716,
+ "loss": 5.270739555358887,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.013468850404024124,
+ "learning_rate": 0.002710882210281016,
+ "loss": 5.27161169052124,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.030262840911746025,
+ "learning_rate": 0.002710091490286264,
+ "loss": 5.256673812866211,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.013560288585722446,
+ "learning_rate": 0.0027092998074278846,
+ "loss": 5.253674507141113,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014592813327908516,
+ "learning_rate": 0.002708507162343721,
+ "loss": 5.25804328918457,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014328169636428356,
+ "learning_rate": 0.0027077135556723953,
+ "loss": 5.235117435455322,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.01277366653084755,
+ "learning_rate": 0.0027069189880533,
+ "loss": 5.2053985595703125,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.012948785908520222,
+ "learning_rate": 0.0027061234601266043,
+ "loss": 5.214653015136719,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.014892828650772572,
+ "learning_rate": 0.002705326972533252,
+ "loss": 5.227547645568848,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01427230704575777,
+ "learning_rate": 0.002704529525914957,
+ "loss": 5.241065502166748,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.01650116965174675,
+ "learning_rate": 0.0027037311209142092,
+ "loss": 5.227713108062744,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.019728578627109528,
+ "learning_rate": 0.0027029317581742677,
+ "loss": 5.2624921798706055,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.023847149685025215,
+ "learning_rate": 0.0027021314383391657,
+ "loss": 5.278811454772949,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.020690469071269035,
+ "learning_rate": 0.0027013301620537056,
+ "loss": 5.238919258117676,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.021211959421634674,
+ "learning_rate": 0.0027005279299634616,
+ "loss": 5.253684997558594,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.025854922831058502,
+ "learning_rate": 0.002699724742714778,
+ "loss": 5.2538652420043945,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.0215485580265522,
+ "learning_rate": 0.0026989206009547678,
+ "loss": 5.224540710449219,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.019827792420983315,
+ "learning_rate": 0.0026981155053313137,
+ "loss": 5.220674514770508,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.017689093947410583,
+ "learning_rate": 0.002697309456493066,
+ "loss": 5.228511810302734,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.022564806044101715,
+ "learning_rate": 0.002696502455089445,
+ "loss": 5.295933723449707,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.020070280879735947,
+ "learning_rate": 0.0026956945017706357,
+ "loss": 5.230920791625977,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.025744661688804626,
+ "learning_rate": 0.002694885597187593,
+ "loss": 5.296453952789307,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.021263007074594498,
+ "learning_rate": 0.0026940757419920357,
+ "loss": 5.220481872558594,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.017577245831489563,
+ "learning_rate": 0.0026932649368364495,
+ "loss": 5.231488227844238,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.015562832355499268,
+ "learning_rate": 0.002692453182374086,
+ "loss": 5.175224304199219,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.014115403406322002,
+ "learning_rate": 0.0026916404792589606,
+ "loss": 5.228387832641602,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.012863830663263798,
+ "learning_rate": 0.002690826828145854,
+ "loss": 5.186727523803711,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.014151126146316528,
+ "learning_rate": 0.002690012229690309,
+ "loss": 5.1694817543029785,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.01316932775080204,
+ "learning_rate": 0.0026891966845486353,
+ "loss": 5.19931697845459,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.011340726166963577,
+ "learning_rate": 0.0026883801933779003,
+ "loss": 5.238619804382324,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.012616089545190334,
+ "learning_rate": 0.0026875627568359378,
+ "loss": 5.183376789093018,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.012274038046598434,
+ "learning_rate": 0.002686744375581341,
+ "loss": 5.199946403503418,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.014990398660302162,
+ "learning_rate": 0.002685925050273465,
+ "loss": 5.168451309204102,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.011399757117033005,
+ "learning_rate": 0.0026851047815724255,
+ "loss": 5.173828125,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.013638390228152275,
+ "learning_rate": 0.0026842835701390982,
+ "loss": 5.142702102661133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.014472718350589275,
+ "learning_rate": 0.0026834614166351177,
+ "loss": 5.173436164855957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.014509808272123337,
+ "learning_rate": 0.002682638321722879,
+ "loss": 5.173975467681885,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.013576583005487919,
+ "learning_rate": 0.0026818142860655352,
+ "loss": 5.1435089111328125,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.012914472259581089,
+ "learning_rate": 0.0026809893103269954,
+ "loss": 5.151381492614746,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.013803352601826191,
+ "learning_rate": 0.0026801633951719288,
+ "loss": 5.133159637451172,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.018488213419914246,
+ "learning_rate": 0.0026793365412657606,
+ "loss": 5.183903694152832,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.01843969337642193,
+ "learning_rate": 0.002678508749274671,
+ "loss": 5.1660356521606445,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.017181839793920517,
+ "learning_rate": 0.0026776800198655983,
+ "loss": 5.179847717285156,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0148160420358181,
+ "learning_rate": 0.0026768503537062333,
+ "loss": 5.149198532104492,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.013251296244561672,
+ "learning_rate": 0.0026760197514650247,
+ "loss": 5.192000389099121,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.010746927931904793,
+ "learning_rate": 0.0026751882138111717,
+ "loss": 5.138680458068848,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.012316994369029999,
+ "learning_rate": 0.002674355741414631,
+ "loss": 5.169965744018555,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.01191797386854887,
+ "learning_rate": 0.0026735223349461106,
+ "loss": 5.128267288208008,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.013525084592401981,
+ "learning_rate": 0.0026726879950770698,
+ "loss": 5.124795436859131,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.014843370765447617,
+ "learning_rate": 0.002671852722479721,
+ "loss": 5.1497087478637695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.015285412780940533,
+ "learning_rate": 0.00267101651782703,
+ "loss": 5.093118190765381,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.013555832207202911,
+ "learning_rate": 0.0026701793817927107,
+ "loss": 5.13249397277832,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.010569742880761623,
+ "learning_rate": 0.0026693413150512273,
+ "loss": 5.084077835083008,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.011483813636004925,
+ "learning_rate": 0.0026685023182777975,
+ "loss": 5.128049850463867,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.013724366202950478,
+ "learning_rate": 0.0026676623921483844,
+ "loss": 5.157539367675781,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.014337432570755482,
+ "learning_rate": 0.0026668215373397016,
+ "loss": 5.149948596954346,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.014101454056799412,
+ "learning_rate": 0.00266597975452921,
+ "loss": 5.117250442504883,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.016004368662834167,
+ "learning_rate": 0.0026651370443951197,
+ "loss": 5.143221855163574,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.249833583831787,
+ "eval_runtime": 51.5543,
+ "eval_samples_per_second": 47.368,
+ "eval_steps_per_second": 1.494,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.01297403872013092,
+ "learning_rate": 0.0026642934076163867,
+ "loss": 5.089287281036377,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.012867718935012817,
+ "learning_rate": 0.0026634488448727144,
+ "loss": 5.0784101486206055,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.014600927010178566,
+ "learning_rate": 0.0026626033568445504,
+ "loss": 5.050220489501953,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.016619572415947914,
+ "learning_rate": 0.0026617569442130905,
+ "loss": 5.079303741455078,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.015560587868094444,
+ "learning_rate": 0.0026609096076602736,
+ "loss": 5.072385311126709,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.01774057187139988,
+ "learning_rate": 0.0026600613478687833,
+ "loss": 5.122529983520508,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.019971400499343872,
+ "learning_rate": 0.0026592121655220472,
+ "loss": 5.085966110229492,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.017667256295681,
+ "learning_rate": 0.0026583620613042365,
+ "loss": 5.078432559967041,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.022369660437107086,
+ "learning_rate": 0.002657511035900264,
+ "loss": 5.0630903244018555,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.017964458093047142,
+ "learning_rate": 0.002656659089995787,
+ "loss": 5.0846076011657715,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.021769145503640175,
+ "learning_rate": 0.002655806224277201,
+ "loss": 5.081344127655029,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.02487620897591114,
+ "learning_rate": 0.0026549524394316464,
+ "loss": 5.106287956237793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.022516218945384026,
+ "learning_rate": 0.0026540977361470012,
+ "loss": 5.128098487854004,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.02099926583468914,
+ "learning_rate": 0.0026532421151118845,
+ "loss": 5.103562355041504,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.01899387687444687,
+ "learning_rate": 0.002652385577015655,
+ "loss": 5.062675476074219,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.016801996156573296,
+ "learning_rate": 0.002651528122548409,
+ "loss": 5.096002578735352,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.014009200967848301,
+ "learning_rate": 0.0026506697524009836,
+ "loss": 5.075201034545898,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.014653858728706837,
+ "learning_rate": 0.002649810467264951,
+ "loss": 5.090478897094727,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.014918150380253792,
+ "learning_rate": 0.0026489502678326226,
+ "loss": 5.070797920227051,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.01327288057655096,
+ "learning_rate": 0.0026480891547970443,
+ "loss": 5.030500411987305,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.012885140255093575,
+ "learning_rate": 0.002647227128852,
+ "loss": 5.0845537185668945,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.012848512269556522,
+ "learning_rate": 0.002646364190692009,
+ "loss": 5.079484939575195,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.01149789709597826,
+ "learning_rate": 0.002645500341012324,
+ "loss": 5.036316394805908,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.010360757820308208,
+ "learning_rate": 0.0026446355805089337,
+ "loss": 5.04129695892334,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.010426290333271027,
+ "learning_rate": 0.00264376990987856,
+ "loss": 5.051270484924316,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.01018153689801693,
+ "learning_rate": 0.002642903329818658,
+ "loss": 5.102342128753662,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.011879283003509045,
+ "learning_rate": 0.0026420358410274153,
+ "loss": 5.053501129150391,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.010418329387903214,
+ "learning_rate": 0.002641167444203752,
+ "loss": 5.030653476715088,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.012302013114094734,
+ "learning_rate": 0.0026402981400473202,
+ "loss": 5.0402069091796875,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.011392916552722454,
+ "learning_rate": 0.002639427929258502,
+ "loss": 5.024111747741699,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.013075058348476887,
+ "learning_rate": 0.0026385568125384114,
+ "loss": 5.021731376647949,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015349832363426685,
+ "learning_rate": 0.0026376847905888903,
+ "loss": 5.03192138671875,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01588309183716774,
+ "learning_rate": 0.0026368118641125113,
+ "loss": 4.966363906860352,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.014378590509295464,
+ "learning_rate": 0.0026359380338125764,
+ "loss": 5.070773124694824,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.012306655757129192,
+ "learning_rate": 0.002635063300393114,
+ "loss": 5.025704860687256,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01212004292756319,
+ "learning_rate": 0.002634187664558881,
+ "loss": 4.965104103088379,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013538849540054798,
+ "learning_rate": 0.0026333111270153624,
+ "loss": 5.029810905456543,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.012052906677126884,
+ "learning_rate": 0.0026324336884687672,
+ "loss": 5.0262908935546875,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.011368660256266594,
+ "learning_rate": 0.002631555349626033,
+ "loss": 5.001485824584961,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.010215395130217075,
+ "learning_rate": 0.0026306761111948205,
+ "loss": 5.019739151000977,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.011449878104031086,
+ "learning_rate": 0.0026297959738835167,
+ "loss": 4.9816083908081055,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.012381811626255512,
+ "learning_rate": 0.0026289149384012326,
+ "loss": 5.025032043457031,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.00974641740322113,
+ "learning_rate": 0.002628033005457802,
+ "loss": 5.000177383422852,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.009515874087810516,
+ "learning_rate": 0.0026271501757637833,
+ "loss": 4.996979713439941,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.011125455610454082,
+ "learning_rate": 0.0026262664500304555,
+ "loss": 4.996796607971191,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.013290468603372574,
+ "learning_rate": 0.0026253818289698207,
+ "loss": 5.053703308105469,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.014650283381342888,
+ "learning_rate": 0.002624496313294602,
+ "loss": 5.001321792602539,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014537827111780643,
+ "learning_rate": 0.002623609903718244,
+ "loss": 5.024584770202637,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.012908566743135452,
+ "learning_rate": 0.0026227226009549105,
+ "loss": 5.042878150939941,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.01446808222681284,
+ "learning_rate": 0.002621834405719485,
+ "loss": 5.020857810974121,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.01535153016448021,
+ "learning_rate": 0.0026209453187275708,
+ "loss": 5.0056304931640625,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01815616339445114,
+ "learning_rate": 0.0026200553406954892,
+ "loss": 5.0467376708984375,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.023135516792535782,
+ "learning_rate": 0.0026191644723402795,
+ "loss": 5.033995628356934,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.019682930782437325,
+ "learning_rate": 0.002618272714379698,
+ "loss": 5.008149147033691,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.018521904945373535,
+ "learning_rate": 0.002617380067532217,
+ "loss": 5.007076263427734,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.02118663489818573,
+ "learning_rate": 0.0026164865325170278,
+ "loss": 5.05519962310791,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024684788659214973,
+ "learning_rate": 0.0026155921100540344,
+ "loss": 5.027571201324463,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.02609819918870926,
+ "learning_rate": 0.0026146968008638567,
+ "loss": 5.10361385345459,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.025958053767681122,
+ "learning_rate": 0.00261380060566783,
+ "loss": 5.0911455154418945,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.017067518085241318,
+ "learning_rate": 0.002612903525188002,
+ "loss": 5.045783042907715,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.016510486602783203,
+ "learning_rate": 0.0026120055601471345,
+ "loss": 5.069464683532715,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015552333556115627,
+ "learning_rate": 0.0026111067112687012,
+ "loss": 5.024248123168945,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.013432685285806656,
+ "learning_rate": 0.0026102069792768886,
+ "loss": 5.004837989807129,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.012300201691687107,
+ "learning_rate": 0.0026093063648965943,
+ "loss": 5.031093597412109,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.01197744719684124,
+ "learning_rate": 0.002608404868853428,
+ "loss": 4.965237617492676,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.01100246049463749,
+ "learning_rate": 0.002607502491873708,
+ "loss": 5.04719352722168,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.010111447423696518,
+ "learning_rate": 0.0026065992346844626,
+ "loss": 4.986410617828369,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.010669790208339691,
+ "learning_rate": 0.002605695098013431,
+ "loss": 4.9825663566589355,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.010805477388203144,
+ "learning_rate": 0.0026047900825890593,
+ "loss": 4.991506576538086,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.010886074975132942,
+ "learning_rate": 0.002603884189140502,
+ "loss": 4.987531661987305,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.010735839605331421,
+ "learning_rate": 0.0026029774183976206,
+ "loss": 4.982756614685059,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.011092593893408775,
+ "learning_rate": 0.002602069771090985,
+ "loss": 4.994824409484863,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 5.13364315032959,
+ "eval_runtime": 50.4195,
+ "eval_samples_per_second": 48.434,
+ "eval_steps_per_second": 1.527,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.01005905494093895,
+ "learning_rate": 0.0026011612479518694,
+ "loss": 4.918412685394287,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.009904682636260986,
+ "learning_rate": 0.002600251849712255,
+ "loss": 4.924630165100098,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.009543103165924549,
+ "learning_rate": 0.0025993415771048273,
+ "loss": 4.940309524536133,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.011031470261514187,
+ "learning_rate": 0.0025984304308629765,
+ "loss": 4.924609184265137,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.009511022828519344,
+ "learning_rate": 0.002597518411720796,
+ "loss": 4.9102325439453125,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.00886621605604887,
+ "learning_rate": 0.0025966055204130847,
+ "loss": 4.9236907958984375,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.010057039558887482,
+ "learning_rate": 0.002595691757675341,
+ "loss": 4.908633232116699,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.01098555512726307,
+ "learning_rate": 0.0025947771242437684,
+ "loss": 4.935927391052246,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.008800056762993336,
+ "learning_rate": 0.0025938616208552694,
+ "loss": 4.873291969299316,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.009747766889631748,
+ "learning_rate": 0.0025929452482474497,
+ "loss": 4.954436302185059,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.010785684920847416,
+ "learning_rate": 0.002592028007158613,
+ "loss": 4.924927711486816,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.012620717287063599,
+ "learning_rate": 0.0025911098983277645,
+ "loss": 4.953606605529785,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.014428738504648209,
+ "learning_rate": 0.002590190922494607,
+ "loss": 4.951706886291504,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.014496048912405968,
+ "learning_rate": 0.0025892710803995444,
+ "loss": 4.941249847412109,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.0109238987788558,
+ "learning_rate": 0.0025883503727836764,
+ "loss": 4.939975261688232,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.012271072715520859,
+ "learning_rate": 0.0025874288003887996,
+ "loss": 4.880942344665527,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.016302887350320816,
+ "learning_rate": 0.0025865063639574086,
+ "loss": 4.931602478027344,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.016137251630425453,
+ "learning_rate": 0.0025855830642326945,
+ "loss": 4.918013095855713,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.014732961542904377,
+ "learning_rate": 0.002584658901958542,
+ "loss": 4.927577495574951,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.016387946903705597,
+ "learning_rate": 0.0025837338778795325,
+ "loss": 4.930819034576416,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.017146151512861252,
+ "learning_rate": 0.002582807992740941,
+ "loss": 4.928377151489258,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.016153613105416298,
+ "learning_rate": 0.0025818812472887375,
+ "loss": 4.975607395172119,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.013386047445237637,
+ "learning_rate": 0.0025809536422695825,
+ "loss": 4.9434709548950195,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.012328253127634525,
+ "learning_rate": 0.0025800251784308314,
+ "loss": 4.917585372924805,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.012000586837530136,
+ "learning_rate": 0.0025790958565205316,
+ "loss": 4.9137725830078125,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.009463997557759285,
+ "learning_rate": 0.002578165677287419,
+ "loss": 4.910362243652344,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.010591360740363598,
+ "learning_rate": 0.0025772346414809242,
+ "loss": 4.916032314300537,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.010429762303829193,
+ "learning_rate": 0.002576302749851164,
+ "loss": 4.934561729431152,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.12526583671569824,
+ "learning_rate": 0.0025753700031489486,
+ "loss": 4.897515773773193,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01368892751634121,
+ "learning_rate": 0.0025744364021257743,
+ "loss": 4.94814920425415,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.016603967174887657,
+ "learning_rate": 0.0025735019475338254,
+ "loss": 4.8996758460998535,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.06110106781125069,
+ "learning_rate": 0.0025725666401259775,
+ "loss": 4.946794509887695,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.018292423337697983,
+ "learning_rate": 0.002571630480655789,
+ "loss": 4.958992958068848,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.01852245442569256,
+ "learning_rate": 0.0025706934698775073,
+ "loss": 4.942556381225586,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.0195649154484272,
+ "learning_rate": 0.0025697556085460646,
+ "loss": 4.895773887634277,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.017820244655013084,
+ "learning_rate": 0.00256881689741708,
+ "loss": 4.915250778198242,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.016814742237329483,
+ "learning_rate": 0.002567877337246855,
+ "loss": 4.9521260261535645,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.015319781377911568,
+ "learning_rate": 0.002566936928792377,
+ "loss": 4.899402141571045,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.01607401855289936,
+ "learning_rate": 0.0025659956728113146,
+ "loss": 4.909913063049316,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.015564955770969391,
+ "learning_rate": 0.002565053570062023,
+ "loss": 4.947719573974609,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.017502037808299065,
+ "learning_rate": 0.002564110621303536,
+ "loss": 4.95107364654541,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.014791073277592659,
+ "learning_rate": 0.002563166827295571,
+ "loss": 4.8995866775512695,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.012700124643743038,
+ "learning_rate": 0.002562222188798525,
+ "loss": 4.922100067138672,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.011235272511839867,
+ "learning_rate": 0.0025612767065734763,
+ "loss": 4.884456634521484,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.013009768910706043,
+ "learning_rate": 0.002560330381382183,
+ "loss": 4.869503021240234,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.0102638378739357,
+ "learning_rate": 0.0025593832139870834,
+ "loss": 4.934551239013672,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.011081870645284653,
+ "learning_rate": 0.0025584352051512913,
+ "loss": 4.937268257141113,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.010681218467652798,
+ "learning_rate": 0.002557486355638601,
+ "loss": 4.920559883117676,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.010962890461087227,
+ "learning_rate": 0.002556536666213484,
+ "loss": 4.936082363128662,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.010790379717946053,
+ "learning_rate": 0.0025555861376410868,
+ "loss": 4.9506449699401855,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.010659157298505306,
+ "learning_rate": 0.0025546347706872346,
+ "loss": 4.872883319854736,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.012962655164301395,
+ "learning_rate": 0.0025536825661184254,
+ "loss": 4.901210784912109,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.014810293912887573,
+ "learning_rate": 0.002552729524701833,
+ "loss": 4.895804405212402,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.010993832722306252,
+ "learning_rate": 0.0025517756472053068,
+ "loss": 4.891035079956055,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.013413915410637856,
+ "learning_rate": 0.002550820934397368,
+ "loss": 4.900317192077637,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.01580147072672844,
+ "learning_rate": 0.0025498653870472117,
+ "loss": 4.883669853210449,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01864512264728546,
+ "learning_rate": 0.0025489090059247045,
+ "loss": 4.881938457489014,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017185743898153305,
+ "learning_rate": 0.0025479517918003862,
+ "loss": 4.853325843811035,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.012045156210660934,
+ "learning_rate": 0.002546993745445466,
+ "loss": 4.9152326583862305,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.013420728035271168,
+ "learning_rate": 0.0025460348676318257,
+ "loss": 4.895130157470703,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.011561610735952854,
+ "learning_rate": 0.0025450751591320152,
+ "loss": 4.890257835388184,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013164008036255836,
+ "learning_rate": 0.0025441146207192547,
+ "loss": 4.885972023010254,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01120204571634531,
+ "learning_rate": 0.002543153253167432,
+ "loss": 4.882246971130371,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.011342627927660942,
+ "learning_rate": 0.002542191057251104,
+ "loss": 4.862927436828613,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.010852213948965073,
+ "learning_rate": 0.0025412280337454942,
+ "loss": 4.879744529724121,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.012851282022893429,
+ "learning_rate": 0.0025402641834264936,
+ "loss": 4.882238388061523,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.011839655227959156,
+ "learning_rate": 0.002539299507070659,
+ "loss": 4.85658073425293,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.011315262876451015,
+ "learning_rate": 0.002538334005455213,
+ "loss": 4.9234771728515625,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.011661105789244175,
+ "learning_rate": 0.0025373676793580416,
+ "loss": 4.855696678161621,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.009712521918118,
+ "learning_rate": 0.0025364005295576975,
+ "loss": 4.89942741394043,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.011002928018569946,
+ "learning_rate": 0.0025354325568333956,
+ "loss": 4.871459007263184,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.01509513333439827,
+ "learning_rate": 0.0025344637619650135,
+ "loss": 4.834778785705566,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 5.031810283660889,
+ "eval_runtime": 51.9412,
+ "eval_samples_per_second": 47.015,
+ "eval_steps_per_second": 1.482,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.010317740961909294,
+ "learning_rate": 0.0025334941457330925,
+ "loss": 4.8169708251953125,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.00892715621739626,
+ "learning_rate": 0.0025325237089188343,
+ "loss": 4.852178573608398,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.011608307249844074,
+ "learning_rate": 0.002531552452304103,
+ "loss": 4.790487289428711,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.013100114651024342,
+ "learning_rate": 0.0025305803766714216,
+ "loss": 4.863536834716797,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.016571057960391045,
+ "learning_rate": 0.002529607482803974,
+ "loss": 4.840987205505371,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.014595622196793556,
+ "learning_rate": 0.002528633771485604,
+ "loss": 4.815027236938477,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.014603931456804276,
+ "learning_rate": 0.002527659243500812,
+ "loss": 4.853024959564209,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.019107041880488396,
+ "learning_rate": 0.0025266838996347595,
+ "loss": 4.829808235168457,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.020292190834879875,
+ "learning_rate": 0.0025257077406732613,
+ "loss": 4.775232315063477,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.014630074612796307,
+ "learning_rate": 0.0025247307674027926,
+ "loss": 4.826682090759277,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.019507668912410736,
+ "learning_rate": 0.0025237529806104823,
+ "loss": 4.843222141265869,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.023950330913066864,
+ "learning_rate": 0.002522774381084116,
+ "loss": 4.820512771606445,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.022547820582985878,
+ "learning_rate": 0.002521794969612133,
+ "loss": 4.854910850524902,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.017819277942180634,
+ "learning_rate": 0.0025208147469836283,
+ "loss": 4.825885772705078,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.017989592626690865,
+ "learning_rate": 0.002519833713988349,
+ "loss": 4.8357014656066895,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.013516640290617943,
+ "learning_rate": 0.0025188518714166956,
+ "loss": 4.83729362487793,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.015956101939082146,
+ "learning_rate": 0.002517869220059721,
+ "loss": 4.807845592498779,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.013967473991215229,
+ "learning_rate": 0.002516885760709131,
+ "loss": 4.828944206237793,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.01270931027829647,
+ "learning_rate": 0.0025159014941572785,
+ "loss": 4.826594352722168,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.012433716095983982,
+ "learning_rate": 0.002514916421197171,
+ "loss": 4.797939777374268,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.011882808059453964,
+ "learning_rate": 0.0025139305426224636,
+ "loss": 4.78917932510376,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.011193005368113518,
+ "learning_rate": 0.0025129438592274614,
+ "loss": 4.787964820861816,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.011873407289385796,
+ "learning_rate": 0.002511956371807116,
+ "loss": 4.81754732131958,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.011407418176531792,
+ "learning_rate": 0.002510968081157029,
+ "loss": 4.781639575958252,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.009560304693877697,
+ "learning_rate": 0.002509978988073449,
+ "loss": 4.8163957595825195,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.009969786740839481,
+ "learning_rate": 0.0025089890933532697,
+ "loss": 4.795045852661133,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.010739767923951149,
+ "learning_rate": 0.0025079983977940312,
+ "loss": 4.847584247589111,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.009999158792197704,
+ "learning_rate": 0.0025070069021939194,
+ "loss": 4.834909439086914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.009302482940256596,
+ "learning_rate": 0.002506014607351765,
+ "loss": 4.808745384216309,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.010687744244933128,
+ "learning_rate": 0.0025050215140670403,
+ "loss": 4.789101600646973,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.010930616408586502,
+ "learning_rate": 0.0025040276231398637,
+ "loss": 4.817464351654053,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.013101065531373024,
+ "learning_rate": 0.0025030329353709956,
+ "loss": 4.827780723571777,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.01223307941108942,
+ "learning_rate": 0.0025020374515618373,
+ "loss": 4.8367414474487305,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.013091634958982468,
+ "learning_rate": 0.0025010411725144323,
+ "loss": 4.839454650878906,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.012114843353629112,
+ "learning_rate": 0.0025000440990314643,
+ "loss": 4.832447052001953,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014664588496088982,
+ "learning_rate": 0.0024990462319162588,
+ "loss": 4.740812301635742,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.015566283836960793,
+ "learning_rate": 0.002498047571972777,
+ "loss": 4.815671920776367,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.012440434657037258,
+ "learning_rate": 0.002497048120005623,
+ "loss": 4.769775867462158,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01175499614328146,
+ "learning_rate": 0.0024960478768200362,
+ "loss": 4.8101348876953125,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.012699508108198643,
+ "learning_rate": 0.002495046843221896,
+ "loss": 4.827281951904297,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.012438303790986538,
+ "learning_rate": 0.002494045020017715,
+ "loss": 4.786289691925049,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.01709924265742302,
+ "learning_rate": 0.002493042408014646,
+ "loss": 4.828742027282715,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.022684535011649132,
+ "learning_rate": 0.0024920390080204733,
+ "loss": 4.822258949279785,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.011984334327280521,
+ "learning_rate": 0.00249103482084362,
+ "loss": 4.792006969451904,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.013941102661192417,
+ "learning_rate": 0.00249002984729314,
+ "loss": 4.7926177978515625,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.0145634226500988,
+ "learning_rate": 0.002489024088178724,
+ "loss": 4.819155216217041,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.013391612097620964,
+ "learning_rate": 0.0024880175443106932,
+ "loss": 4.777242660522461,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01764756441116333,
+ "learning_rate": 0.0024870102165000015,
+ "loss": 4.81627893447876,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.015397441573441029,
+ "learning_rate": 0.002486002105558235,
+ "loss": 4.869495391845703,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01225620973855257,
+ "learning_rate": 0.0024849932122976105,
+ "loss": 4.784395217895508,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.01534759160131216,
+ "learning_rate": 0.002483983537530975,
+ "loss": 4.813475608825684,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.015057779848575592,
+ "learning_rate": 0.002482973082071805,
+ "loss": 4.8281569480896,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.01243986189365387,
+ "learning_rate": 0.0024819618467342066,
+ "loss": 4.7866973876953125,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.013326527550816536,
+ "learning_rate": 0.0024809498323329136,
+ "loss": 4.750214576721191,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.013644695281982422,
+ "learning_rate": 0.002479937039683288,
+ "loss": 4.790144920349121,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01296213548630476,
+ "learning_rate": 0.002478923469601319,
+ "loss": 4.812170028686523,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.011848201975226402,
+ "learning_rate": 0.0024779091229036204,
+ "loss": 4.798855781555176,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.012765968218445778,
+ "learning_rate": 0.0024768940004074335,
+ "loss": 4.8272294998168945,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.011700666509568691,
+ "learning_rate": 0.002475878102930625,
+ "loss": 4.762265205383301,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.01018498931080103,
+ "learning_rate": 0.0024748614312916843,
+ "loss": 4.796628952026367,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.010241681709885597,
+ "learning_rate": 0.002473843986309725,
+ "loss": 4.829319000244141,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.009332080371677876,
+ "learning_rate": 0.0024728257688044855,
+ "loss": 4.783773899078369,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.012860726565122604,
+ "learning_rate": 0.002471806779596324,
+ "loss": 4.802700519561768,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.011156697757542133,
+ "learning_rate": 0.0024707870195062225,
+ "loss": 4.808038234710693,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.010329720564186573,
+ "learning_rate": 0.002469766489355783,
+ "loss": 4.761645317077637,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.011093895882368088,
+ "learning_rate": 0.0024687451899672278,
+ "loss": 4.795220851898193,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.010740782134234905,
+ "learning_rate": 0.0024677231221634,
+ "loss": 4.79289436340332,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.012272508814930916,
+ "learning_rate": 0.0024667002867677606,
+ "loss": 4.811873435974121,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.008665879257023335,
+ "learning_rate": 0.0024656766846043905,
+ "loss": 4.744179725646973,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.00933373998850584,
+ "learning_rate": 0.0024646523164979866,
+ "loss": 4.805453777313232,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.010399337857961655,
+ "learning_rate": 0.002463627183273864,
+ "loss": 4.75568962097168,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.013210986740887165,
+ "learning_rate": 0.0024626012857579542,
+ "loss": 4.797974586486816,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.967026233673096,
+ "eval_runtime": 51.07,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.01731858029961586,
+ "learning_rate": 0.0024615746247768043,
+ "loss": 4.748838901519775,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.018902689218521118,
+ "learning_rate": 0.0024605472011575763,
+ "loss": 4.717338562011719,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.01734415628015995,
+ "learning_rate": 0.0024595190157280468,
+ "loss": 4.737737655639648,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.01971321739256382,
+ "learning_rate": 0.002458490069316607,
+ "loss": 4.774558067321777,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.020029844716191292,
+ "learning_rate": 0.002457460362752259,
+ "loss": 4.739579200744629,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.020853109657764435,
+ "learning_rate": 0.0024564298968646203,
+ "loss": 4.791436672210693,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.019159914925694466,
+ "learning_rate": 0.0024553986724839176,
+ "loss": 4.765683650970459,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.0165473110973835,
+ "learning_rate": 0.0024543666904409903,
+ "loss": 4.732603073120117,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.017426826059818268,
+ "learning_rate": 0.002453333951567287,
+ "loss": 4.7360639572143555,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014590734615921974,
+ "learning_rate": 0.002452300456694868,
+ "loss": 4.750031471252441,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.011172222904860973,
+ "learning_rate": 0.0024512662066563996,
+ "loss": 4.740771293640137,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.012294141575694084,
+ "learning_rate": 0.0024502312022851606,
+ "loss": 4.726214408874512,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.01146282535046339,
+ "learning_rate": 0.002449195444415034,
+ "loss": 4.753599166870117,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.011197247542440891,
+ "learning_rate": 0.0024481589338805113,
+ "loss": 4.751479148864746,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.013252820819616318,
+ "learning_rate": 0.0024471216715166904,
+ "loss": 4.743409156799316,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.012366211041808128,
+ "learning_rate": 0.0024460836581592744,
+ "loss": 4.731625556945801,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.013906655833125114,
+ "learning_rate": 0.0024450448946445737,
+ "loss": 4.768153190612793,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.012118754908442497,
+ "learning_rate": 0.002444005381809499,
+ "loss": 4.74198579788208,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.01197550818324089,
+ "learning_rate": 0.002442965120491568,
+ "loss": 4.737289905548096,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.010640372522175312,
+ "learning_rate": 0.002441924111528901,
+ "loss": 4.7472686767578125,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.013553396798670292,
+ "learning_rate": 0.0024408823557602196,
+ "loss": 4.711215972900391,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.01393711194396019,
+ "learning_rate": 0.002439839854024847,
+ "loss": 4.720060348510742,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01589774526655674,
+ "learning_rate": 0.002438796607162709,
+ "loss": 4.695697784423828,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.0097908740863204,
+ "learning_rate": 0.0024377526160143303,
+ "loss": 4.729333877563477,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.011745489202439785,
+ "learning_rate": 0.0024367078814208363,
+ "loss": 4.722957134246826,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.009649662300944328,
+ "learning_rate": 0.0024356624042239498,
+ "loss": 4.727540493011475,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.008429240435361862,
+ "learning_rate": 0.002434616185265993,
+ "loss": 4.705052375793457,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.00984937883913517,
+ "learning_rate": 0.002433569225389887,
+ "loss": 4.721161365509033,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.009237290360033512,
+ "learning_rate": 0.0024325215254391466,
+ "loss": 4.724277496337891,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.008900545537471771,
+ "learning_rate": 0.002431473086257886,
+ "loss": 4.711522102355957,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.010003026574850082,
+ "learning_rate": 0.002430423908690813,
+ "loss": 4.728816032409668,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.011658324860036373,
+ "learning_rate": 0.0024293739935832316,
+ "loss": 4.716179847717285,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.01312185451388359,
+ "learning_rate": 0.0024283233417810393,
+ "loss": 4.7172346115112305,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.014306646771728992,
+ "learning_rate": 0.0024272719541307266,
+ "loss": 4.704054832458496,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.009844634681940079,
+ "learning_rate": 0.0024262198314793783,
+ "loss": 4.72379207611084,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.00903896801173687,
+ "learning_rate": 0.0024251669746746703,
+ "loss": 4.711025238037109,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.010476433672010899,
+ "learning_rate": 0.0024241133845648703,
+ "loss": 4.711664199829102,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.009128391742706299,
+ "learning_rate": 0.0024230590619988375,
+ "loss": 4.72395133972168,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.01228832732886076,
+ "learning_rate": 0.0024220040078260195,
+ "loss": 4.747347831726074,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.011916602961719036,
+ "learning_rate": 0.002420948222896455,
+ "loss": 4.696471691131592,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.012877700850367546,
+ "learning_rate": 0.0024198917080607706,
+ "loss": 4.731119632720947,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.012967764399945736,
+ "learning_rate": 0.0024188344641701814,
+ "loss": 4.688953876495361,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.011993546038866043,
+ "learning_rate": 0.002417776492076489,
+ "loss": 4.695237159729004,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.011188571341335773,
+ "learning_rate": 0.0024167177926320833,
+ "loss": 4.699850082397461,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.010556348599493504,
+ "learning_rate": 0.0024156583666899395,
+ "loss": 4.703063011169434,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.012023619376122952,
+ "learning_rate": 0.0024145982151036176,
+ "loss": 4.6918792724609375,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01266001258045435,
+ "learning_rate": 0.0024135373387272623,
+ "loss": 4.6653289794921875,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.012152725830674171,
+ "learning_rate": 0.0024124757384156024,
+ "loss": 4.682897567749023,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.014056944288313389,
+ "learning_rate": 0.002411413415023951,
+ "loss": 4.696136474609375,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.016826556995511055,
+ "learning_rate": 0.002410350369408202,
+ "loss": 4.701657295227051,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.015048800967633724,
+ "learning_rate": 0.0024092866024248323,
+ "loss": 4.691389083862305,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.012512968853116035,
+ "learning_rate": 0.0024082221149308997,
+ "loss": 4.677886009216309,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01596425101161003,
+ "learning_rate": 0.0024071569077840425,
+ "loss": 4.718337535858154,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01640196517109871,
+ "learning_rate": 0.002406090981842479,
+ "loss": 4.697303295135498,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.012347170151770115,
+ "learning_rate": 0.002405024337965007,
+ "loss": 4.696938991546631,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01213028747588396,
+ "learning_rate": 0.0024039569770110012,
+ "loss": 4.737306594848633,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01314542256295681,
+ "learning_rate": 0.002402888899840416,
+ "loss": 4.69948673248291,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.013764189556241035,
+ "learning_rate": 0.0024018201073137805,
+ "loss": 4.722800254821777,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.011147351004183292,
+ "learning_rate": 0.002400750600292203,
+ "loss": 4.694892883300781,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.015437614172697067,
+ "learning_rate": 0.0023996803796373653,
+ "loss": 4.704752445220947,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.016402097418904305,
+ "learning_rate": 0.0023986094462115255,
+ "loss": 4.67849063873291,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.011600594967603683,
+ "learning_rate": 0.002397537800877514,
+ "loss": 4.689641952514648,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.011864298023283482,
+ "learning_rate": 0.0023964654444987373,
+ "loss": 4.670377731323242,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01409872155636549,
+ "learning_rate": 0.002395392377939173,
+ "loss": 4.696022033691406,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.012106554582715034,
+ "learning_rate": 0.0023943186020633717,
+ "loss": 4.705615043640137,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.012310667894780636,
+ "learning_rate": 0.002393244117736455,
+ "loss": 4.702986717224121,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.013501073233783245,
+ "learning_rate": 0.002392168925824116,
+ "loss": 4.6700215339660645,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.013565126806497574,
+ "learning_rate": 0.0023910930271926166,
+ "loss": 4.699914455413818,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.013421077281236649,
+ "learning_rate": 0.00239001642270879,
+ "loss": 4.713191986083984,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.014408102259039879,
+ "learning_rate": 0.002388939113240035,
+ "loss": 4.655015468597412,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.013502835296094418,
+ "learning_rate": 0.0023878610996543224,
+ "loss": 4.6142258644104,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.014169570058584213,
+ "learning_rate": 0.002386782382820188,
+ "loss": 4.660243034362793,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.885881423950195,
+ "eval_runtime": 52.0315,
+ "eval_samples_per_second": 46.933,
+ "eval_steps_per_second": 1.48,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.014618083834648132,
+ "learning_rate": 0.002385702963606734,
+ "loss": 4.640046119689941,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01745622046291828,
+ "learning_rate": 0.0023846228428836278,
+ "loss": 4.641966819763184,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.014995258301496506,
+ "learning_rate": 0.002383542021521106,
+ "loss": 4.60922908782959,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.013278020545840263,
+ "learning_rate": 0.002382460500389964,
+ "loss": 4.652071952819824,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.014685257337987423,
+ "learning_rate": 0.002381378280361566,
+ "loss": 4.672174453735352,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.012737024575471878,
+ "learning_rate": 0.002380295362307836,
+ "loss": 4.612064361572266,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.012992842122912407,
+ "learning_rate": 0.002379211747101261,
+ "loss": 4.611995697021484,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.010210646316409111,
+ "learning_rate": 0.0023781274356148912,
+ "loss": 4.6055707931518555,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.010265499353408813,
+ "learning_rate": 0.0023770424287223365,
+ "loss": 4.637089729309082,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.010174009948968887,
+ "learning_rate": 0.002375956727297768,
+ "loss": 4.620571136474609,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.009512446820735931,
+ "learning_rate": 0.0023748703322159143,
+ "loss": 4.644092559814453,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.009849252179265022,
+ "learning_rate": 0.0023737832443520655,
+ "loss": 4.625055313110352,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.010493188165128231,
+ "learning_rate": 0.0023726954645820674,
+ "loss": 4.644410610198975,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.022181199863553047,
+ "learning_rate": 0.002371606993782326,
+ "loss": 4.593084335327148,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.011906919069588184,
+ "learning_rate": 0.0023705178328298016,
+ "loss": 4.6452836990356445,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.011874495074152946,
+ "learning_rate": 0.0023694279826020122,
+ "loss": 4.599322319030762,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.00989708211272955,
+ "learning_rate": 0.0023683374439770295,
+ "loss": 4.607342720031738,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.009256849065423012,
+ "learning_rate": 0.0023672462178334817,
+ "loss": 4.621583461761475,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.013187132775783539,
+ "learning_rate": 0.0023661543050505496,
+ "loss": 4.614629745483398,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.013505950570106506,
+ "learning_rate": 0.002365061706507968,
+ "loss": 4.629693984985352,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.012032733298838139,
+ "learning_rate": 0.0023639684230860225,
+ "loss": 4.671209335327148,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.01278337650001049,
+ "learning_rate": 0.002362874455665553,
+ "loss": 4.648874759674072,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.01123690977692604,
+ "learning_rate": 0.0023617798051279496,
+ "loss": 4.611470699310303,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.012564311735332012,
+ "learning_rate": 0.002360684472355151,
+ "loss": 4.659026145935059,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.01206638477742672,
+ "learning_rate": 0.002359588458229649,
+ "loss": 4.642566204071045,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.010300020687282085,
+ "learning_rate": 0.0023584917636344806,
+ "loss": 4.629585266113281,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.0102265365421772,
+ "learning_rate": 0.0023573943894532336,
+ "loss": 4.609250068664551,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.012440572492778301,
+ "learning_rate": 0.002356296336570042,
+ "loss": 4.636714458465576,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.011830241419374943,
+ "learning_rate": 0.0023551976058695887,
+ "loss": 4.59100866317749,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.011482699774205685,
+ "learning_rate": 0.0023540981982370997,
+ "loss": 4.620088577270508,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.012950294651091099,
+ "learning_rate": 0.0023529981145583483,
+ "loss": 4.578657627105713,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.012197710573673248,
+ "learning_rate": 0.002351897355719652,
+ "loss": 4.620813846588135,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.014839230105280876,
+ "learning_rate": 0.002350795922607872,
+ "loss": 4.615731239318848,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016670791432261467,
+ "learning_rate": 0.002349693816110414,
+ "loss": 4.655855178833008,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.014426100067794323,
+ "learning_rate": 0.0023485910371152247,
+ "loss": 4.660849571228027,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.014040560461580753,
+ "learning_rate": 0.0023474875865107936,
+ "loss": 4.634917736053467,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018269391730427742,
+ "learning_rate": 0.0023463834651861496,
+ "loss": 4.611809253692627,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.013855630531907082,
+ "learning_rate": 0.0023452786740308647,
+ "loss": 4.615225791931152,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.011505477130413055,
+ "learning_rate": 0.0023441732139350487,
+ "loss": 4.62592887878418,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.010166114196181297,
+ "learning_rate": 0.002343067085789351,
+ "loss": 4.6219024658203125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.009253021329641342,
+ "learning_rate": 0.0023419602904849596,
+ "loss": 4.658320426940918,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01027526892721653,
+ "learning_rate": 0.002340852828913598,
+ "loss": 4.589777946472168,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.008855022490024567,
+ "learning_rate": 0.002339744701967529,
+ "loss": 4.588204383850098,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.00965463649481535,
+ "learning_rate": 0.002338635910539551,
+ "loss": 4.648177146911621,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.009073720313608646,
+ "learning_rate": 0.0023375264555229956,
+ "loss": 4.585582733154297,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.010105383582413197,
+ "learning_rate": 0.0023364163378117317,
+ "loss": 4.612174987792969,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.015307556837797165,
+ "learning_rate": 0.002335305558300161,
+ "loss": 4.591984748840332,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016716254875063896,
+ "learning_rate": 0.002334194117883219,
+ "loss": 4.659285545349121,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.013281598687171936,
+ "learning_rate": 0.0023330820174563725,
+ "loss": 4.645829200744629,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.014416513964533806,
+ "learning_rate": 0.0023319692579156204,
+ "loss": 4.584893226623535,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.016326071694493294,
+ "learning_rate": 0.002330855840157494,
+ "loss": 4.633223533630371,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.012851824052631855,
+ "learning_rate": 0.002329741765079053,
+ "loss": 4.580986499786377,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.015444506891071796,
+ "learning_rate": 0.0023286270335778876,
+ "loss": 4.615891456604004,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.01391979493200779,
+ "learning_rate": 0.0023275116465521174,
+ "loss": 4.628403186798096,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.012138686142861843,
+ "learning_rate": 0.0023263956049003895,
+ "loss": 4.6259942054748535,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.013741672039031982,
+ "learning_rate": 0.0023252789095218785,
+ "loss": 4.6056742668151855,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01103319600224495,
+ "learning_rate": 0.002324161561316285,
+ "loss": 4.621670246124268,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.010333316400647163,
+ "learning_rate": 0.0023230435611838367,
+ "loss": 4.620484352111816,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.011063606478273869,
+ "learning_rate": 0.0023219249100252864,
+ "loss": 4.64952278137207,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.009984857402741909,
+ "learning_rate": 0.0023208056087419106,
+ "loss": 4.572571754455566,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.009986856952309608,
+ "learning_rate": 0.002319685658235511,
+ "loss": 4.570366382598877,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.010905452072620392,
+ "learning_rate": 0.0023185650594084103,
+ "loss": 4.603614807128906,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.012800061143934727,
+ "learning_rate": 0.002317443813163456,
+ "loss": 4.562705039978027,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.01252608560025692,
+ "learning_rate": 0.002316321920404015,
+ "loss": 4.599016189575195,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.011305592954158783,
+ "learning_rate": 0.002315199382033976,
+ "loss": 4.644617557525635,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.010594413615763187,
+ "learning_rate": 0.002314076198957749,
+ "loss": 4.569808006286621,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.009859138168394566,
+ "learning_rate": 0.002312952372080261,
+ "loss": 4.5890398025512695,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.010982917621731758,
+ "learning_rate": 0.00231182790230696,
+ "loss": 4.606348991394043,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.010252755135297775,
+ "learning_rate": 0.00231070279054381,
+ "loss": 4.58942174911499,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01110257301479578,
+ "learning_rate": 0.0023095770376972947,
+ "loss": 4.599823474884033,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01003815233707428,
+ "learning_rate": 0.0023084506446744103,
+ "loss": 4.569343090057373,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.010517258197069168,
+ "learning_rate": 0.0023073236123826745,
+ "loss": 4.646301746368408,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.804611682891846,
+ "eval_runtime": 51.6279,
+ "eval_samples_per_second": 47.3,
+ "eval_steps_per_second": 1.491,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.011439066380262375,
+ "learning_rate": 0.0023061959417301145,
+ "loss": 4.563335418701172,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.01073157787322998,
+ "learning_rate": 0.002305067633625275,
+ "loss": 4.540966987609863,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.011391228064894676,
+ "learning_rate": 0.0023039386889772136,
+ "loss": 4.555583953857422,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.012814595364034176,
+ "learning_rate": 0.0023028091086955,
+ "loss": 4.517425537109375,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.012994988821446896,
+ "learning_rate": 0.002301678893690218,
+ "loss": 4.538477897644043,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.01460991520434618,
+ "learning_rate": 0.00230054804487196,
+ "loss": 4.571145057678223,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.013036075979471207,
+ "learning_rate": 0.0022994165631518306,
+ "loss": 4.519233703613281,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.011917057447135448,
+ "learning_rate": 0.0022982844494414444,
+ "loss": 4.5071587562561035,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.013928866013884544,
+ "learning_rate": 0.0022971517046529257,
+ "loss": 4.531422138214111,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01407038513571024,
+ "learning_rate": 0.0022960183296989053,
+ "loss": 4.511205196380615,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.019396567717194557,
+ "learning_rate": 0.0022948843254925245,
+ "loss": 4.530007362365723,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.014425835572183132,
+ "learning_rate": 0.002293749692947428,
+ "loss": 4.530138969421387,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.01600247249007225,
+ "learning_rate": 0.00229261443297777,
+ "loss": 4.5643391609191895,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.018238037824630737,
+ "learning_rate": 0.0022914785464982095,
+ "loss": 4.540779113769531,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.015427535399794579,
+ "learning_rate": 0.0022903420344239085,
+ "loss": 4.506287097930908,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.016816625371575356,
+ "learning_rate": 0.0022892048976705347,
+ "loss": 4.537919521331787,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.01999717950820923,
+ "learning_rate": 0.002288067137154259,
+ "loss": 4.558426380157471,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.020884918048977852,
+ "learning_rate": 0.002286928753791753,
+ "loss": 4.592915058135986,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.01620141603052616,
+ "learning_rate": 0.002285789748500194,
+ "loss": 4.553356170654297,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.018495604395866394,
+ "learning_rate": 0.0022846501221972567,
+ "loss": 4.573139190673828,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.0156320258975029,
+ "learning_rate": 0.0022835098758011174,
+ "loss": 4.585921764373779,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.014515317976474762,
+ "learning_rate": 0.002282369010230452,
+ "loss": 4.5532002449035645,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.014826556667685509,
+ "learning_rate": 0.0022812275264044355,
+ "loss": 4.560998916625977,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.013240831904113293,
+ "learning_rate": 0.00228008542524274,
+ "loss": 4.531787872314453,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.013814719393849373,
+ "learning_rate": 0.0022789427076655372,
+ "loss": 4.611872673034668,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.012246561236679554,
+ "learning_rate": 0.0022777993745934934,
+ "loss": 4.5565385818481445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01206218171864748,
+ "learning_rate": 0.002276655426947771,
+ "loss": 4.570852279663086,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01007329672574997,
+ "learning_rate": 0.0022755108656500275,
+ "loss": 4.553400039672852,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.010361720807850361,
+ "learning_rate": 0.0022743656916224173,
+ "loss": 4.527509689331055,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.009456256404519081,
+ "learning_rate": 0.0022732199057875846,
+ "loss": 4.520236015319824,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.009310049936175346,
+ "learning_rate": 0.002272073509068669,
+ "loss": 4.518343925476074,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.00927449855953455,
+ "learning_rate": 0.0022709265023893015,
+ "loss": 4.559690475463867,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.008658594451844692,
+ "learning_rate": 0.002269778886673605,
+ "loss": 4.513876914978027,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.008928220719099045,
+ "learning_rate": 0.0022686306628461918,
+ "loss": 4.545050621032715,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.008046358823776245,
+ "learning_rate": 0.002267481831832166,
+ "loss": 4.54866361618042,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.008318924345076084,
+ "learning_rate": 0.00226633239455712,
+ "loss": 4.546968936920166,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.008808869868516922,
+ "learning_rate": 0.0022651823519471343,
+ "loss": 4.50877571105957,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.008767383173108101,
+ "learning_rate": 0.0022640317049287774,
+ "loss": 4.57666540145874,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.009351354092359543,
+ "learning_rate": 0.0022628804544291047,
+ "loss": 4.511702537536621,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.010137057863175869,
+ "learning_rate": 0.002261728601375659,
+ "loss": 4.539676189422607,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.00882046204060316,
+ "learning_rate": 0.002260576146696466,
+ "loss": 4.533196449279785,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.009354782290756702,
+ "learning_rate": 0.0022594230913200388,
+ "loss": 4.515488624572754,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.009241201914846897,
+ "learning_rate": 0.002258269436175372,
+ "loss": 4.542706489562988,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.009252829477190971,
+ "learning_rate": 0.002257115182191947,
+ "loss": 4.592660903930664,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.011243131943047047,
+ "learning_rate": 0.0022559603302997225,
+ "loss": 4.527804851531982,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.011032823473215103,
+ "learning_rate": 0.0022548048814291437,
+ "loss": 4.540122032165527,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.010495275259017944,
+ "learning_rate": 0.0022536488365111347,
+ "loss": 4.532835006713867,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.009992090053856373,
+ "learning_rate": 0.0022524921964770994,
+ "loss": 4.521217346191406,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01141198631376028,
+ "learning_rate": 0.0022513349622589225,
+ "loss": 4.519709587097168,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.013087889179587364,
+ "learning_rate": 0.002250177134788966,
+ "loss": 4.547289848327637,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.01328477542847395,
+ "learning_rate": 0.0022490187150000706,
+ "loss": 4.527865409851074,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01033794041723013,
+ "learning_rate": 0.002247859703825555,
+ "loss": 4.54740047454834,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.010522300377488136,
+ "learning_rate": 0.002246700102199213,
+ "loss": 4.538840293884277,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.012151090428233147,
+ "learning_rate": 0.0022455399110553145,
+ "loss": 4.536865234375,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.010553102008998394,
+ "learning_rate": 0.0022443791313286054,
+ "loss": 4.5348334312438965,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.011433709412813187,
+ "learning_rate": 0.0022432177639543035,
+ "loss": 4.510526657104492,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01189691387116909,
+ "learning_rate": 0.002242055809868103,
+ "loss": 4.530807018280029,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.010031254962086678,
+ "learning_rate": 0.0022408932700061683,
+ "loss": 4.496850967407227,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.010129331611096859,
+ "learning_rate": 0.002239730145305138,
+ "loss": 4.551287651062012,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.010430586524307728,
+ "learning_rate": 0.0022385664367021185,
+ "loss": 4.542672157287598,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.011774476617574692,
+ "learning_rate": 0.002237402145134691,
+ "loss": 4.537034034729004,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.01184659544378519,
+ "learning_rate": 0.002236237271540903,
+ "loss": 4.521193504333496,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.011327247135341167,
+ "learning_rate": 0.0022350718168592724,
+ "loss": 4.49426794052124,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.010767918080091476,
+ "learning_rate": 0.0022339057820287855,
+ "loss": 4.531225204467773,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.011195424944162369,
+ "learning_rate": 0.002232739167988894,
+ "loss": 4.492671012878418,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.010394908487796783,
+ "learning_rate": 0.0022315719756795197,
+ "loss": 4.523616790771484,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.010277998633682728,
+ "learning_rate": 0.002230404206041047,
+ "loss": 4.487832069396973,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.010955783538520336,
+ "learning_rate": 0.0022292358600143275,
+ "loss": 4.526865005493164,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.010611076839268208,
+ "learning_rate": 0.002228066938540676,
+ "loss": 4.498682022094727,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.010008828714489937,
+ "learning_rate": 0.0022268974425618724,
+ "loss": 4.524019241333008,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.008973540738224983,
+ "learning_rate": 0.002225727373020157,
+ "loss": 4.512996673583984,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.011754963546991348,
+ "learning_rate": 0.002224556730858235,
+ "loss": 4.496376991271973,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.744190216064453,
+ "eval_runtime": 51.6572,
+ "eval_samples_per_second": 47.273,
+ "eval_steps_per_second": 1.491,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.012094628997147083,
+ "learning_rate": 0.0022233855170192703,
+ "loss": 4.443447113037109,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.014347990974783897,
+ "learning_rate": 0.00222221373244689,
+ "loss": 4.485690116882324,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.021286984905600548,
+ "learning_rate": 0.0022210413780851786,
+ "loss": 4.487505912780762,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.020273910835385323,
+ "learning_rate": 0.0022198684548786827,
+ "loss": 4.473119735717773,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.017642971128225327,
+ "learning_rate": 0.002218694963772403,
+ "loss": 4.440615653991699,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.02075147069990635,
+ "learning_rate": 0.0022175209057118017,
+ "loss": 4.527026176452637,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016479508951306343,
+ "learning_rate": 0.002216346281642796,
+ "loss": 4.508321762084961,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.017978593707084656,
+ "learning_rate": 0.002215171092511759,
+ "loss": 4.500985145568848,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.01739897020161152,
+ "learning_rate": 0.002213995339265519,
+ "loss": 4.474225044250488,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.015556776896119118,
+ "learning_rate": 0.0022128190228513597,
+ "loss": 4.444849014282227,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.016166578978300095,
+ "learning_rate": 0.002211642144217018,
+ "loss": 4.5026726722717285,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.01534261740744114,
+ "learning_rate": 0.0022104647043106832,
+ "loss": 4.484527587890625,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.013517106883227825,
+ "learning_rate": 0.002209286704080998,
+ "loss": 4.453907012939453,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.012562787160277367,
+ "learning_rate": 0.0022081081444770556,
+ "loss": 4.471528053283691,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.009947682730853558,
+ "learning_rate": 0.0022069290264483997,
+ "loss": 4.504153728485107,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01169571466743946,
+ "learning_rate": 0.002205749350945026,
+ "loss": 4.447345733642578,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.011394352652132511,
+ "learning_rate": 0.0022045691189173755,
+ "loss": 4.447491645812988,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.011168593540787697,
+ "learning_rate": 0.0022033883313163417,
+ "loss": 4.441639423370361,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.009503341279923916,
+ "learning_rate": 0.002202206989093263,
+ "loss": 4.461601257324219,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.0112199941650033,
+ "learning_rate": 0.002201025093199926,
+ "loss": 4.444095134735107,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.010505333542823792,
+ "learning_rate": 0.002199842644588562,
+ "loss": 4.460987091064453,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.009494813159108162,
+ "learning_rate": 0.0021986596442118496,
+ "loss": 4.460158348083496,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.01063452661037445,
+ "learning_rate": 0.00219747609302291,
+ "loss": 4.489625453948975,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.012194888666272163,
+ "learning_rate": 0.0021962919919753087,
+ "loss": 4.441605567932129,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.010613647289574146,
+ "learning_rate": 0.0021951073420230557,
+ "loss": 4.475821018218994,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.009643700905144215,
+ "learning_rate": 0.002193922144120601,
+ "loss": 4.467228889465332,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.00912613607943058,
+ "learning_rate": 0.0021927363992228373,
+ "loss": 4.455435752868652,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.009371735155582428,
+ "learning_rate": 0.0021915501082850988,
+ "loss": 4.479915142059326,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.008617875166237354,
+ "learning_rate": 0.002190363272263158,
+ "loss": 4.415642738342285,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.009571383707225323,
+ "learning_rate": 0.002189175892113227,
+ "loss": 4.439854621887207,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.00782635249197483,
+ "learning_rate": 0.002187987968791957,
+ "loss": 4.484311103820801,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.008408057503402233,
+ "learning_rate": 0.0021867995032564358,
+ "loss": 4.46577262878418,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.009641910903155804,
+ "learning_rate": 0.00218561049646419,
+ "loss": 4.437408447265625,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.008498182520270348,
+ "learning_rate": 0.0021844209493731784,
+ "loss": 4.442599296569824,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.009289798326790333,
+ "learning_rate": 0.0021832308629418,
+ "loss": 4.517641067504883,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.009360183961689472,
+ "learning_rate": 0.0021820402381288848,
+ "loss": 4.474143981933594,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.008764738216996193,
+ "learning_rate": 0.0021808490758936983,
+ "loss": 4.4376702308654785,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.009673701599240303,
+ "learning_rate": 0.0021796573771959375,
+ "loss": 4.435220718383789,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.010852573439478874,
+ "learning_rate": 0.0021784651429957334,
+ "loss": 4.465187072753906,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.011292314156889915,
+ "learning_rate": 0.0021772723742536464,
+ "loss": 4.458176136016846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01071334257721901,
+ "learning_rate": 0.0021760790719306702,
+ "loss": 4.425498962402344,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.01505239401012659,
+ "learning_rate": 0.0021748852369882265,
+ "loss": 4.463129043579102,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.011774689890444279,
+ "learning_rate": 0.0021736908703881663,
+ "loss": 4.462577819824219,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.011241118423640728,
+ "learning_rate": 0.002172495973092769,
+ "loss": 4.452583312988281,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.01072910986840725,
+ "learning_rate": 0.002171300546064742,
+ "loss": 4.466248035430908,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01021308358758688,
+ "learning_rate": 0.002170104590267219,
+ "loss": 4.485857963562012,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.012630444951355457,
+ "learning_rate": 0.00216890810666376,
+ "loss": 4.442112922668457,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.013602569699287415,
+ "learning_rate": 0.0021677110962183515,
+ "loss": 4.45171594619751,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.013808641582727432,
+ "learning_rate": 0.0021665135598954014,
+ "loss": 4.468818187713623,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.011510699987411499,
+ "learning_rate": 0.002165315498659744,
+ "loss": 4.439584255218506,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.010502424091100693,
+ "learning_rate": 0.0021641169134766353,
+ "loss": 4.478377342224121,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.009953407570719719,
+ "learning_rate": 0.0021629178053117542,
+ "loss": 4.474404335021973,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.011158930137753487,
+ "learning_rate": 0.0021617181751312,
+ "loss": 4.4549560546875,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.010431940667331219,
+ "learning_rate": 0.002160518023901493,
+ "loss": 4.444911003112793,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.009100335650146008,
+ "learning_rate": 0.0021593173525895734,
+ "loss": 4.428071975708008,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.009362847544252872,
+ "learning_rate": 0.0021581161621628013,
+ "loss": 4.462827682495117,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.009893129579722881,
+ "learning_rate": 0.0021569144535889535,
+ "loss": 4.426912307739258,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.010524468496441841,
+ "learning_rate": 0.0021557122278362244,
+ "loss": 4.435606002807617,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.010621445253491402,
+ "learning_rate": 0.002154509485873226,
+ "loss": 4.454901695251465,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.010836896486580372,
+ "learning_rate": 0.0021533062286689855,
+ "loss": 4.4441728591918945,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.011403030715882778,
+ "learning_rate": 0.0021521024571929457,
+ "loss": 4.483884811401367,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.012522426433861256,
+ "learning_rate": 0.0021508981724149643,
+ "loss": 4.464006423950195,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.014799179509282112,
+ "learning_rate": 0.0021496933753053105,
+ "loss": 4.460402011871338,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.012453878298401833,
+ "learning_rate": 0.0021484880668346685,
+ "loss": 4.456899642944336,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.012270379811525345,
+ "learning_rate": 0.002147282247974133,
+ "loss": 4.477325439453125,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01094560045748949,
+ "learning_rate": 0.0021460759196952105,
+ "loss": 4.422362327575684,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.009995176456868649,
+ "learning_rate": 0.0021448690829698183,
+ "loss": 4.405127048492432,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.014935536310076714,
+ "learning_rate": 0.002143661738770283,
+ "loss": 4.445840358734131,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.011492030695080757,
+ "learning_rate": 0.0021424538880693393,
+ "loss": 4.459555625915527,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.008793608285486698,
+ "learning_rate": 0.002141245531840132,
+ "loss": 4.436631202697754,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.008894025348126888,
+ "learning_rate": 0.0021400366710562103,
+ "loss": 4.473924160003662,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.010324863716959953,
+ "learning_rate": 0.0021388273066915327,
+ "loss": 4.475680351257324,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.688324451446533,
+ "eval_runtime": 54.4515,
+ "eval_samples_per_second": 44.847,
+ "eval_steps_per_second": 1.414,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.009235628880560398,
+ "learning_rate": 0.0021376174397204615,
+ "loss": 4.370064735412598,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.008818200789391994,
+ "learning_rate": 0.0021364070711177643,
+ "loss": 4.371678352355957,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.010333757847547531,
+ "learning_rate": 0.0021351962018586147,
+ "loss": 4.3896379470825195,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.011938809417188168,
+ "learning_rate": 0.002133984832918587,
+ "loss": 4.361343860626221,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.010634614154696465,
+ "learning_rate": 0.00213277296527366,
+ "loss": 4.362449645996094,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.008545687422156334,
+ "learning_rate": 0.0021315605999002126,
+ "loss": 4.3685503005981445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.009159795008599758,
+ "learning_rate": 0.002130347737775027,
+ "loss": 4.379969596862793,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.010464153252542019,
+ "learning_rate": 0.0021291343798752835,
+ "loss": 4.376113414764404,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.011010527610778809,
+ "learning_rate": 0.0021279205271785622,
+ "loss": 4.390711784362793,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.010350867174565792,
+ "learning_rate": 0.002126706180662843,
+ "loss": 4.399683952331543,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.009708237834274769,
+ "learning_rate": 0.002125491341306503,
+ "loss": 4.363435745239258,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.010087115690112114,
+ "learning_rate": 0.002124276010088316,
+ "loss": 4.366623878479004,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.009712349623441696,
+ "learning_rate": 0.002123060187987453,
+ "loss": 4.360814571380615,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.010174986906349659,
+ "learning_rate": 0.0021218438759834785,
+ "loss": 4.347475051879883,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.010195785202085972,
+ "learning_rate": 0.0021206270750563543,
+ "loss": 4.399251461029053,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.010786959901452065,
+ "learning_rate": 0.0021194097861864354,
+ "loss": 4.397933006286621,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.011476189829409122,
+ "learning_rate": 0.002118192010354468,
+ "loss": 4.3748369216918945,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.011571013368666172,
+ "learning_rate": 0.0021169737485415927,
+ "loss": 4.388349533081055,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.011140935122966766,
+ "learning_rate": 0.002115755001729341,
+ "loss": 4.398018836975098,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.013216734863817692,
+ "learning_rate": 0.0021145357708996365,
+ "loss": 4.406947135925293,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.013385126367211342,
+ "learning_rate": 0.0021133160570347893,
+ "loss": 4.41457462310791,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.011784231290221214,
+ "learning_rate": 0.002112095861117502,
+ "loss": 4.403835296630859,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.009813498705625534,
+ "learning_rate": 0.0021108751841308645,
+ "loss": 4.3865647315979,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.010297936387360096,
+ "learning_rate": 0.0021096540270583533,
+ "loss": 4.395583152770996,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.011965147219598293,
+ "learning_rate": 0.002108432390883834,
+ "loss": 4.385443687438965,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.012212193571031094,
+ "learning_rate": 0.0021072102765915556,
+ "loss": 4.422321319580078,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.009840975515544415,
+ "learning_rate": 0.0021059876851661544,
+ "loss": 4.3568902015686035,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.013039806857705116,
+ "learning_rate": 0.002104764617592649,
+ "loss": 4.314867973327637,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.015091902576386929,
+ "learning_rate": 0.0021035410748564444,
+ "loss": 4.392338752746582,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.015448705293238163,
+ "learning_rate": 0.0021023170579433257,
+ "loss": 4.379609107971191,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.018393104895949364,
+ "learning_rate": 0.0021010925678394622,
+ "loss": 4.418291091918945,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.01866372860968113,
+ "learning_rate": 0.002099867605531402,
+ "loss": 4.4006452560424805,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.01819106750190258,
+ "learning_rate": 0.002098642172006077,
+ "loss": 4.433975696563721,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.019774602726101875,
+ "learning_rate": 0.0020974162682507952,
+ "loss": 4.417821884155273,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.016885802149772644,
+ "learning_rate": 0.0020961898952532457,
+ "loss": 4.397073268890381,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.015750881284475327,
+ "learning_rate": 0.0020949630540014956,
+ "loss": 4.4356608390808105,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.015696514397859573,
+ "learning_rate": 0.002093735745483988,
+ "loss": 4.422449588775635,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.017195021733641624,
+ "learning_rate": 0.0020925079706895435,
+ "loss": 4.443209648132324,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.015864048153162003,
+ "learning_rate": 0.0020912797306073577,
+ "loss": 4.4340691566467285,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.009857800789177418,
+ "learning_rate": 0.0020900510262270022,
+ "loss": 4.37935209274292,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.013739476911723614,
+ "learning_rate": 0.0020888218585384216,
+ "loss": 4.412148475646973,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.011667690239846706,
+ "learning_rate": 0.0020875922285319344,
+ "loss": 4.391700744628906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.012323460541665554,
+ "learning_rate": 0.0020863621371982302,
+ "loss": 4.423880577087402,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.010353626683354378,
+ "learning_rate": 0.0020851315855283724,
+ "loss": 4.427382469177246,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.0101751908659935,
+ "learning_rate": 0.002083900574513794,
+ "loss": 4.415317535400391,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.01126549206674099,
+ "learning_rate": 0.002082669105146298,
+ "loss": 4.402498245239258,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.011096738278865814,
+ "learning_rate": 0.0020814371784180575,
+ "loss": 4.433567047119141,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.010944445617496967,
+ "learning_rate": 0.0020802047953216137,
+ "loss": 4.406990051269531,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.008965318091213703,
+ "learning_rate": 0.0020789719568498747,
+ "loss": 4.3721771240234375,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.00828113779425621,
+ "learning_rate": 0.002077738663996117,
+ "loss": 4.389828205108643,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.008599454537034035,
+ "learning_rate": 0.002076504917753982,
+ "loss": 4.411396026611328,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.007713889703154564,
+ "learning_rate": 0.0020752707191174762,
+ "loss": 4.355343341827393,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.00912355538457632,
+ "learning_rate": 0.0020740360690809723,
+ "loss": 4.423413276672363,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.008269775658845901,
+ "learning_rate": 0.0020728009686392045,
+ "loss": 4.390920639038086,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.007461601868271828,
+ "learning_rate": 0.002071565418787272,
+ "loss": 4.365501403808594,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.007951333187520504,
+ "learning_rate": 0.0020703294205206336,
+ "loss": 4.378884315490723,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.007816503755748272,
+ "learning_rate": 0.002069092974835112,
+ "loss": 4.4241743087768555,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.00833756010979414,
+ "learning_rate": 0.0020678560827268886,
+ "loss": 4.391102313995361,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.008241684176027775,
+ "learning_rate": 0.0020666187451925055,
+ "loss": 4.374970436096191,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.008245996199548244,
+ "learning_rate": 0.002065380963228862,
+ "loss": 4.380222320556641,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.007378706708550453,
+ "learning_rate": 0.0020641427378332183,
+ "loss": 4.391451358795166,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.007594958413392305,
+ "learning_rate": 0.002062904070003189,
+ "loss": 4.372774124145508,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.0076752156019210815,
+ "learning_rate": 0.002061664960736747,
+ "loss": 4.384668827056885,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.008727089501917362,
+ "learning_rate": 0.00206042541103222,
+ "loss": 4.363401412963867,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.007616396062076092,
+ "learning_rate": 0.002059185421888291,
+ "loss": 4.413188934326172,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.008922954089939594,
+ "learning_rate": 0.0020579449943039967,
+ "loss": 4.383417129516602,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.009063739329576492,
+ "learning_rate": 0.002056704129278727,
+ "loss": 4.385283470153809,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.008912574499845505,
+ "learning_rate": 0.002055462827812225,
+ "loss": 4.378478050231934,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.008317525498569012,
+ "learning_rate": 0.0020542210909045844,
+ "loss": 4.367859840393066,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.009347227402031422,
+ "learning_rate": 0.0020529789195562506,
+ "loss": 4.38238000869751,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.011030065827071667,
+ "learning_rate": 0.0020517363147680183,
+ "loss": 4.372624397277832,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.011519700288772583,
+ "learning_rate": 0.002050493277541032,
+ "loss": 4.3906121253967285,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.641542434692383,
+ "eval_runtime": 51.6342,
+ "eval_samples_per_second": 47.294,
+ "eval_steps_per_second": 1.491,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.009694516658782959,
+ "learning_rate": 0.0020492498088767836,
+ "loss": 4.32808256149292,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.0097952950745821,
+ "learning_rate": 0.002048005909777114,
+ "loss": 4.280096054077148,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.010393359698355198,
+ "learning_rate": 0.0020467615812442106,
+ "loss": 4.317824363708496,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.00977372657507658,
+ "learning_rate": 0.0020455168242806054,
+ "loss": 4.3445329666137695,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.008568719029426575,
+ "learning_rate": 0.0020442716398891775,
+ "loss": 4.318026542663574,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.009732510894536972,
+ "learning_rate": 0.0020430260290731487,
+ "loss": 4.330609321594238,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.010429969057440758,
+ "learning_rate": 0.0020417799928360864,
+ "loss": 4.316794395446777,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.009292383678257465,
+ "learning_rate": 0.002040533532181898,
+ "loss": 4.317420959472656,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.011222455650568008,
+ "learning_rate": 0.0020392866481148356,
+ "loss": 4.292610168457031,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.013591725379228592,
+ "learning_rate": 0.0020380393416394894,
+ "loss": 4.325689315795898,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.013572026044130325,
+ "learning_rate": 0.002036791613760794,
+ "loss": 4.334960460662842,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.012986521236598492,
+ "learning_rate": 0.0020355434654840197,
+ "loss": 4.33829402923584,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.013321424834430218,
+ "learning_rate": 0.0020342948978147777,
+ "loss": 4.2926926612854,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.012307215481996536,
+ "learning_rate": 0.0020330459117590153,
+ "loss": 4.32757568359375,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.012909360229969025,
+ "learning_rate": 0.0020317965083230197,
+ "loss": 4.362889289855957,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017746610566973686,
+ "learning_rate": 0.0020305466885134115,
+ "loss": 4.3426313400268555,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.01696564070880413,
+ "learning_rate": 0.002029296453337148,
+ "loss": 4.3298516273498535,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.013348255306482315,
+ "learning_rate": 0.002028045803801522,
+ "loss": 4.360220432281494,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.013225145637989044,
+ "learning_rate": 0.002026794740914158,
+ "loss": 4.346250057220459,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.01569020375609398,
+ "learning_rate": 0.002025543265683016,
+ "loss": 4.308025360107422,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.018612321466207504,
+ "learning_rate": 0.002024291379116385,
+ "loss": 4.330190181732178,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.01619943603873253,
+ "learning_rate": 0.00202303908222289,
+ "loss": 4.339552402496338,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.015385213308036327,
+ "learning_rate": 0.0020217863760114826,
+ "loss": 4.3412184715271,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.018517112359404564,
+ "learning_rate": 0.002020533261491446,
+ "loss": 4.3289313316345215,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.014868509024381638,
+ "learning_rate": 0.0020192797396723903,
+ "loss": 4.359846115112305,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.014240248128771782,
+ "learning_rate": 0.0020180258115642578,
+ "loss": 4.369707107543945,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.018113158643245697,
+ "learning_rate": 0.0020167714781773143,
+ "loss": 4.362668514251709,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.01802460104227066,
+ "learning_rate": 0.002015516740522154,
+ "loss": 4.344742298126221,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.01414549071341753,
+ "learning_rate": 0.0020142615996096964,
+ "loss": 4.358080863952637,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.012157700955867767,
+ "learning_rate": 0.002013006056451186,
+ "loss": 4.347763538360596,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.011268534697592258,
+ "learning_rate": 0.0020117501120581905,
+ "loss": 4.335859298706055,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.010395628400146961,
+ "learning_rate": 0.002010493767442602,
+ "loss": 4.341609001159668,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.012147950939834118,
+ "learning_rate": 0.002009237023616635,
+ "loss": 4.328502655029297,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.00945813488215208,
+ "learning_rate": 0.0020079798815928252,
+ "loss": 4.311186790466309,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.009489500895142555,
+ "learning_rate": 0.0020067223423840294,
+ "loss": 4.371470928192139,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.009360029362142086,
+ "learning_rate": 0.0020054644070034234,
+ "loss": 4.33941125869751,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.010032779537141323,
+ "learning_rate": 0.0020042060764645037,
+ "loss": 4.365797519683838,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.009367046877741814,
+ "learning_rate": 0.002002947351781084,
+ "loss": 4.29443359375,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.008369733579456806,
+ "learning_rate": 0.0020016882339672967,
+ "loss": 4.294233322143555,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.010394785553216934,
+ "learning_rate": 0.0020004287240375893,
+ "loss": 4.352355003356934,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01068816613405943,
+ "learning_rate": 0.001999168823006728,
+ "loss": 4.342516899108887,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.009975973516702652,
+ "learning_rate": 0.0019979085318897905,
+ "loss": 4.346763610839844,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.009618151932954788,
+ "learning_rate": 0.001996647851702171,
+ "loss": 4.329115867614746,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.008399694226682186,
+ "learning_rate": 0.001995386783459577,
+ "loss": 4.28736686706543,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.008235268294811249,
+ "learning_rate": 0.001994125328178029,
+ "loss": 4.339872360229492,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.00836547277867794,
+ "learning_rate": 0.0019928634868738577,
+ "loss": 4.328156471252441,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.009619626216590405,
+ "learning_rate": 0.001991601260563706,
+ "loss": 4.337394714355469,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.009814038872718811,
+ "learning_rate": 0.0019903386502645275,
+ "loss": 4.325926780700684,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.009588954970240593,
+ "learning_rate": 0.001989075656993584,
+ "loss": 4.286477088928223,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.010471943765878677,
+ "learning_rate": 0.001987812281768447,
+ "loss": 4.322466850280762,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.010239035822451115,
+ "learning_rate": 0.001986548525606994,
+ "loss": 4.3487653732299805,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.008935568854212761,
+ "learning_rate": 0.001985284389527411,
+ "loss": 4.303601264953613,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.008624759502708912,
+ "learning_rate": 0.00198401987454819,
+ "loss": 4.34036922454834,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.009296547621488571,
+ "learning_rate": 0.0019827549816881277,
+ "loss": 4.273357391357422,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.009071138687431812,
+ "learning_rate": 0.0019814897119663248,
+ "loss": 4.321051597595215,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.008341389708220959,
+ "learning_rate": 0.0019802240664021876,
+ "loss": 4.343242645263672,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.007881715893745422,
+ "learning_rate": 0.0019789580460154227,
+ "loss": 4.32067346572876,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.008750553242862225,
+ "learning_rate": 0.0019776916518260404,
+ "loss": 4.330263614654541,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.008591460064053535,
+ "learning_rate": 0.001976424884854351,
+ "loss": 4.311957359313965,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.008084543980658054,
+ "learning_rate": 0.001975157746120966,
+ "loss": 4.331319808959961,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.008036594837903976,
+ "learning_rate": 0.001973890236646797,
+ "loss": 4.382656097412109,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.00805772002786398,
+ "learning_rate": 0.0019726223574530527,
+ "loss": 4.297247886657715,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.009019420482218266,
+ "learning_rate": 0.0019713541095612405,
+ "loss": 4.326526165008545,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.009953883476555347,
+ "learning_rate": 0.0019700854939931653,
+ "loss": 4.327356338500977,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.009340492077171803,
+ "learning_rate": 0.0019688165117709273,
+ "loss": 4.343993186950684,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.010395721532404423,
+ "learning_rate": 0.0019675471639169226,
+ "loss": 4.284934997558594,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.009946230798959732,
+ "learning_rate": 0.0019662774514538415,
+ "loss": 4.288936614990234,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.010981100611388683,
+ "learning_rate": 0.0019650073754046695,
+ "loss": 4.339935779571533,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.009688783437013626,
+ "learning_rate": 0.0019637369367926827,
+ "loss": 4.381425857543945,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.009442255832254887,
+ "learning_rate": 0.001962466136641452,
+ "loss": 4.304185390472412,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.009371291846036911,
+ "learning_rate": 0.0019611949759748364,
+ "loss": 4.306120872497559,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.009544814005494118,
+ "learning_rate": 0.001959923455816988,
+ "loss": 4.326251983642578,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.602745532989502,
+ "eval_runtime": 53.5041,
+ "eval_samples_per_second": 45.641,
+ "eval_steps_per_second": 1.439,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.01010364294052124,
+ "learning_rate": 0.0019586515771923474,
+ "loss": 4.24491024017334,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.012722729705274105,
+ "learning_rate": 0.0019573793411256446,
+ "loss": 4.257739543914795,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.01307204831391573,
+ "learning_rate": 0.0019561067486418967,
+ "loss": 4.291684150695801,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.011215966194868088,
+ "learning_rate": 0.0019548338007664098,
+ "loss": 4.237947463989258,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.012943019159138203,
+ "learning_rate": 0.0019535604985247727,
+ "loss": 4.242859363555908,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.01337282732129097,
+ "learning_rate": 0.0019522868429428636,
+ "loss": 4.238934516906738,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.013697991147637367,
+ "learning_rate": 0.0019510128350468438,
+ "loss": 4.285760402679443,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.012182159349322319,
+ "learning_rate": 0.001949738475863158,
+ "loss": 4.290044784545898,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.012620339170098305,
+ "learning_rate": 0.0019484637664185346,
+ "loss": 4.267187118530273,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.011284958571195602,
+ "learning_rate": 0.0019471887077399837,
+ "loss": 4.26662540435791,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.01161450706422329,
+ "learning_rate": 0.001945913300854797,
+ "loss": 4.283192157745361,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.010022968985140324,
+ "learning_rate": 0.001944637546790548,
+ "loss": 4.295843124389648,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.010341878980398178,
+ "learning_rate": 0.001943361446575087,
+ "loss": 4.295636177062988,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.011549185030162334,
+ "learning_rate": 0.0019420850012365457,
+ "loss": 4.288768768310547,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.012518810108304024,
+ "learning_rate": 0.0019408082118033331,
+ "loss": 4.229338645935059,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.012035328894853592,
+ "learning_rate": 0.0019395310793041353,
+ "loss": 4.258639335632324,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.011193782091140747,
+ "learning_rate": 0.0019382536047679146,
+ "loss": 4.239264488220215,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.009930136613547802,
+ "learning_rate": 0.0019369757892239096,
+ "loss": 4.252286434173584,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.00953229982405901,
+ "learning_rate": 0.0019356976337016327,
+ "loss": 4.293435096740723,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.010618340224027634,
+ "learning_rate": 0.0019344191392308716,
+ "loss": 4.235588550567627,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.008806651458144188,
+ "learning_rate": 0.0019331403068416856,
+ "loss": 4.26716423034668,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.008934796787798405,
+ "learning_rate": 0.0019318611375644065,
+ "loss": 4.258462905883789,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.009283283725380898,
+ "learning_rate": 0.0019305816324296386,
+ "loss": 4.257868766784668,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.009224812500178814,
+ "learning_rate": 0.0019293017924682555,
+ "loss": 4.258486747741699,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.008957969956099987,
+ "learning_rate": 0.0019280216187114016,
+ "loss": 4.309563636779785,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.010284051299095154,
+ "learning_rate": 0.0019267411121904894,
+ "loss": 4.275271415710449,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.009167762473225594,
+ "learning_rate": 0.0019254602739372001,
+ "loss": 4.238452911376953,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.009411636739969254,
+ "learning_rate": 0.0019241791049834816,
+ "loss": 4.2735595703125,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.009044171310961246,
+ "learning_rate": 0.0019228976063615497,
+ "loss": 4.27017879486084,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.008840068243443966,
+ "learning_rate": 0.0019216157791038834,
+ "loss": 4.28773307800293,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.009512359276413918,
+ "learning_rate": 0.001920333624243228,
+ "loss": 4.266008377075195,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.009946667589247227,
+ "learning_rate": 0.001919051142812593,
+ "loss": 4.279062271118164,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.00899894442409277,
+ "learning_rate": 0.0019177683358452505,
+ "loss": 4.277679443359375,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.00909668579697609,
+ "learning_rate": 0.0019164852043747338,
+ "loss": 4.311397552490234,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.009787327609956264,
+ "learning_rate": 0.0019152017494348408,
+ "loss": 4.30875825881958,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.008983899839222431,
+ "learning_rate": 0.001913917972059626,
+ "loss": 4.259649276733398,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.008411582559347153,
+ "learning_rate": 0.0019126338732834071,
+ "loss": 4.259811878204346,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.009162918664515018,
+ "learning_rate": 0.0019113494541407579,
+ "loss": 4.271515846252441,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.011304561980068684,
+ "learning_rate": 0.0019100647156665125,
+ "loss": 4.269217491149902,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.012483394704759121,
+ "learning_rate": 0.0019087796588957617,
+ "loss": 4.2949299812316895,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.012091380544006824,
+ "learning_rate": 0.0019074942848638522,
+ "loss": 4.289752006530762,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.010011567734181881,
+ "learning_rate": 0.0019062085946063869,
+ "loss": 4.294095993041992,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.008766217157244682,
+ "learning_rate": 0.0019049225891592226,
+ "loss": 4.256425857543945,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.012862488627433777,
+ "learning_rate": 0.001903636269558471,
+ "loss": 4.287847995758057,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.013917063362896442,
+ "learning_rate": 0.0019023496368404972,
+ "loss": 4.282784938812256,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.012129023671150208,
+ "learning_rate": 0.0019010626920419173,
+ "loss": 4.2486419677734375,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.010348426178097725,
+ "learning_rate": 0.0018997754361995986,
+ "loss": 4.289836883544922,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.011821693740785122,
+ "learning_rate": 0.001898487870350662,
+ "loss": 4.214344024658203,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.013136724010109901,
+ "learning_rate": 0.0018971999955324742,
+ "loss": 4.24208402633667,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.013173168525099754,
+ "learning_rate": 0.0018959118127826538,
+ "loss": 4.270164489746094,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.011018159799277782,
+ "learning_rate": 0.0018946233231390656,
+ "loss": 4.2571916580200195,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.01018658559769392,
+ "learning_rate": 0.0018933345276398227,
+ "loss": 4.283267498016357,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.009412688203155994,
+ "learning_rate": 0.0018920454273232848,
+ "loss": 4.272198677062988,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.007799267303198576,
+ "learning_rate": 0.0018907560232280557,
+ "loss": 4.260125637054443,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.008747464045882225,
+ "learning_rate": 0.0018894663163929863,
+ "loss": 4.2539215087890625,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.008003434166312218,
+ "learning_rate": 0.0018881763078571689,
+ "loss": 4.245024681091309,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.007593846879899502,
+ "learning_rate": 0.0018868859986599415,
+ "loss": 4.237929344177246,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.008099420927464962,
+ "learning_rate": 0.0018855953898408815,
+ "loss": 4.2789106369018555,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.008270486257970333,
+ "learning_rate": 0.0018843044824398097,
+ "loss": 4.263522148132324,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.009260379709303379,
+ "learning_rate": 0.0018830132774967866,
+ "loss": 4.275835037231445,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.009289492852985859,
+ "learning_rate": 0.0018817217760521132,
+ "loss": 4.313778400421143,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.009493921883404255,
+ "learning_rate": 0.0018804299791463281,
+ "loss": 4.303672790527344,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.009020023047924042,
+ "learning_rate": 0.0018791378878202105,
+ "loss": 4.245473384857178,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.009446556679904461,
+ "learning_rate": 0.0018778455031147724,
+ "loss": 4.267327785491943,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.009499974548816681,
+ "learning_rate": 0.0018765528260712672,
+ "loss": 4.301205635070801,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.008399393409490585,
+ "learning_rate": 0.00187525985773118,
+ "loss": 4.307802677154541,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.009519701823592186,
+ "learning_rate": 0.0018739665991362327,
+ "loss": 4.284796714782715,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.01125232595950365,
+ "learning_rate": 0.0018726730513283804,
+ "loss": 4.285930633544922,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.011608537286520004,
+ "learning_rate": 0.00187137921534981,
+ "loss": 4.257689476013184,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.01064533181488514,
+ "learning_rate": 0.001870085092242944,
+ "loss": 4.249063968658447,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.009854599833488464,
+ "learning_rate": 0.0018687906830504321,
+ "loss": 4.255159378051758,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.01121739111840725,
+ "learning_rate": 0.0018674959888151576,
+ "loss": 4.287777900695801,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.5675368309021,
+ "eval_runtime": 51.8555,
+ "eval_samples_per_second": 47.092,
+ "eval_steps_per_second": 1.485,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.0104464590549469,
+ "learning_rate": 0.0018662010105802314,
+ "loss": 4.206080436706543,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.012800181284546852,
+ "learning_rate": 0.0018649057493889945,
+ "loss": 4.213550567626953,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.013382901437580585,
+ "learning_rate": 0.001863610206285015,
+ "loss": 4.190858840942383,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.014347050338983536,
+ "learning_rate": 0.0018623143823120894,
+ "loss": 4.22884464263916,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.01347152329981327,
+ "learning_rate": 0.001861018278514239,
+ "loss": 4.215579032897949,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.012786377221345901,
+ "learning_rate": 0.0018597218959357112,
+ "loss": 4.174999237060547,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.014137202873826027,
+ "learning_rate": 0.001858425235620979,
+ "loss": 4.21709680557251,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.013864399865269661,
+ "learning_rate": 0.001857128298614737,
+ "loss": 4.2375030517578125,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.010397658683359623,
+ "learning_rate": 0.0018558310859619048,
+ "loss": 4.231938362121582,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.00888819433748722,
+ "learning_rate": 0.0018545335987076223,
+ "loss": 4.195562839508057,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.010698067955672741,
+ "learning_rate": 0.001853235837897252,
+ "loss": 4.22847318649292,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.00987589918076992,
+ "learning_rate": 0.0018519378045763766,
+ "loss": 4.2147040367126465,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.011451782658696175,
+ "learning_rate": 0.001850639499790798,
+ "loss": 4.179263114929199,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.011224704794585705,
+ "learning_rate": 0.0018493409245865368,
+ "loss": 4.23030948638916,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.00996861606836319,
+ "learning_rate": 0.0018480420800098313,
+ "loss": 4.194283962249756,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.012143119238317013,
+ "learning_rate": 0.0018467429671071374,
+ "loss": 4.221704959869385,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.013126783072948456,
+ "learning_rate": 0.0018454435869251268,
+ "loss": 4.187748908996582,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.013031953014433384,
+ "learning_rate": 0.001844143940510687,
+ "loss": 4.248250961303711,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.012868993915617466,
+ "learning_rate": 0.001842844028910919,
+ "loss": 4.20161247253418,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.010102041065692902,
+ "learning_rate": 0.0018415438531731395,
+ "loss": 4.213654518127441,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.009599820710718632,
+ "learning_rate": 0.0018402434143448747,
+ "loss": 4.206313133239746,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.011536866426467896,
+ "learning_rate": 0.0018389427134738658,
+ "loss": 4.237072944641113,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.009558089077472687,
+ "learning_rate": 0.0018376417516080636,
+ "loss": 4.182195663452148,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01056219544261694,
+ "learning_rate": 0.0018363405297956302,
+ "loss": 4.232585906982422,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.012673584744334221,
+ "learning_rate": 0.0018350390490849354,
+ "loss": 4.228830337524414,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.011215884238481522,
+ "learning_rate": 0.0018337373105245598,
+ "loss": 4.216056823730469,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.011281915940344334,
+ "learning_rate": 0.00183243531516329,
+ "loss": 4.233404159545898,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.010381227359175682,
+ "learning_rate": 0.0018311330640501204,
+ "loss": 4.182013034820557,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.00900198519229889,
+ "learning_rate": 0.0018298305582342513,
+ "loss": 4.215488433837891,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.010198601521551609,
+ "learning_rate": 0.0018285277987650876,
+ "loss": 4.19389533996582,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.009965402074158192,
+ "learning_rate": 0.0018272247866922401,
+ "loss": 4.218005657196045,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.009299935773015022,
+ "learning_rate": 0.0018259215230655202,
+ "loss": 4.223684310913086,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.008977951481938362,
+ "learning_rate": 0.0018246180089349462,
+ "loss": 4.215778350830078,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.009261307306587696,
+ "learning_rate": 0.0018233142453507344,
+ "loss": 4.215066909790039,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.008322088047862053,
+ "learning_rate": 0.0018220102333633044,
+ "loss": 4.2188005447387695,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.008716144599020481,
+ "learning_rate": 0.0018207059740232739,
+ "loss": 4.187867641448975,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.008897020481526852,
+ "learning_rate": 0.0018194014683814625,
+ "loss": 4.218034267425537,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.007298174314200878,
+ "learning_rate": 0.0018180967174888857,
+ "loss": 4.2017011642456055,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.008491789922118187,
+ "learning_rate": 0.0018167917223967582,
+ "loss": 4.239873886108398,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.008413378149271011,
+ "learning_rate": 0.0018154864841564908,
+ "loss": 4.234771251678467,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.008141471073031425,
+ "learning_rate": 0.0018141810038196909,
+ "loss": 4.185944080352783,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.009480944834649563,
+ "learning_rate": 0.0018128752824381603,
+ "loss": 4.206586837768555,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.0106047997251153,
+ "learning_rate": 0.001811569321063894,
+ "loss": 4.235287666320801,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.010663514956831932,
+ "learning_rate": 0.0018102631207490838,
+ "loss": 4.204899311065674,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.011473523452877998,
+ "learning_rate": 0.001808956682546109,
+ "loss": 4.253668308258057,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01165992021560669,
+ "learning_rate": 0.0018076500075075456,
+ "loss": 4.209845542907715,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.011692410334944725,
+ "learning_rate": 0.0018063430966861564,
+ "loss": 4.250368595123291,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01061506848782301,
+ "learning_rate": 0.0018050359511348965,
+ "loss": 4.193821907043457,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.010510941036045551,
+ "learning_rate": 0.0018037285719069097,
+ "loss": 4.254327774047852,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.011587041430175304,
+ "learning_rate": 0.001802420960055528,
+ "loss": 4.2838945388793945,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.01257420051842928,
+ "learning_rate": 0.0018011131166342693,
+ "loss": 4.258241653442383,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.00936233438551426,
+ "learning_rate": 0.0017998050426968407,
+ "loss": 4.201854705810547,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.010093127377331257,
+ "learning_rate": 0.0017984967392971338,
+ "loss": 4.193051338195801,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.010032009333372116,
+ "learning_rate": 0.0017971882074892236,
+ "loss": 4.230212211608887,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.009455587714910507,
+ "learning_rate": 0.0017958794483273726,
+ "loss": 4.2157673835754395,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.009549238719046116,
+ "learning_rate": 0.001794570462866023,
+ "loss": 4.260272026062012,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.011059699580073357,
+ "learning_rate": 0.0017932612521598013,
+ "loss": 4.214818000793457,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.011653627268970013,
+ "learning_rate": 0.0017919518172635146,
+ "loss": 4.205112457275391,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.012438314035534859,
+ "learning_rate": 0.001790642159232151,
+ "loss": 4.179045677185059,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.011234673671424389,
+ "learning_rate": 0.0017893322791208784,
+ "loss": 4.239317417144775,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.010923845693469048,
+ "learning_rate": 0.0017880221779850434,
+ "loss": 4.230571269989014,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.009219309315085411,
+ "learning_rate": 0.001786711856880171,
+ "loss": 4.229068756103516,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.009849398396909237,
+ "learning_rate": 0.0017854013168619638,
+ "loss": 4.254772186279297,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.01117763202637434,
+ "learning_rate": 0.001784090558986299,
+ "loss": 4.252647399902344,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.010664612054824829,
+ "learning_rate": 0.0017827795843092316,
+ "loss": 4.204596519470215,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.00847950391471386,
+ "learning_rate": 0.0017814683938869895,
+ "loss": 4.211061954498291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.007305726408958435,
+ "learning_rate": 0.0017801569887759753,
+ "loss": 4.181221008300781,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.007970666512846947,
+ "learning_rate": 0.0017788453700327653,
+ "loss": 4.237855434417725,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.008873614482581615,
+ "learning_rate": 0.0017775335387141054,
+ "loss": 4.229188919067383,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.009691286832094193,
+ "learning_rate": 0.0017762214958769156,
+ "loss": 4.205491542816162,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.009360672906041145,
+ "learning_rate": 0.0017749092425782852,
+ "loss": 4.233466625213623,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.009895884431898594,
+ "learning_rate": 0.0017735967798754723,
+ "loss": 4.187763214111328,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.535672187805176,
+ "eval_runtime": 53.1736,
+ "eval_samples_per_second": 45.925,
+ "eval_steps_per_second": 1.448,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.00966129545122385,
+ "learning_rate": 0.0017722841088259045,
+ "loss": 4.1451735496521,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.010733380913734436,
+ "learning_rate": 0.0017709712304871774,
+ "loss": 4.1616363525390625,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.0117328567430377,
+ "learning_rate": 0.0017696581459170535,
+ "loss": 4.126331806182861,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.012640298344194889,
+ "learning_rate": 0.0017683448561734607,
+ "loss": 4.163730621337891,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.013074076734483242,
+ "learning_rate": 0.0017670313623144934,
+ "loss": 4.199502944946289,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.01132719125598669,
+ "learning_rate": 0.0017657176653984097,
+ "loss": 4.143380165100098,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.010008333250880241,
+ "learning_rate": 0.0017644037664836319,
+ "loss": 4.170320510864258,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.011090578511357307,
+ "learning_rate": 0.0017630896666287438,
+ "loss": 4.1571478843688965,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.011047456413507462,
+ "learning_rate": 0.0017617753668924924,
+ "loss": 4.101892948150635,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.010098653845489025,
+ "learning_rate": 0.0017604608683337857,
+ "loss": 4.154077529907227,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.010499860160052776,
+ "learning_rate": 0.00175914617201169,
+ "loss": 4.116868019104004,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.010939318686723709,
+ "learning_rate": 0.0017578312789854338,
+ "loss": 4.159144878387451,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.010622999630868435,
+ "learning_rate": 0.0017565161903144023,
+ "loss": 4.171411991119385,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.0108700692653656,
+ "learning_rate": 0.0017552009070581386,
+ "loss": 4.100748062133789,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.009794988669455051,
+ "learning_rate": 0.0017538854302763423,
+ "loss": 4.185161590576172,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.007750354241579771,
+ "learning_rate": 0.0017525697610288702,
+ "loss": 4.149083137512207,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.009923134930431843,
+ "learning_rate": 0.001751253900375732,
+ "loss": 4.205166816711426,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.010811141692101955,
+ "learning_rate": 0.0017499378493770936,
+ "loss": 4.131865978240967,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.010868688113987446,
+ "learning_rate": 0.0017486216090932733,
+ "loss": 4.148791790008545,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.011224126443266869,
+ "learning_rate": 0.001747305180584743,
+ "loss": 4.152503490447998,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.013262554071843624,
+ "learning_rate": 0.001745988564912124,
+ "loss": 4.164926528930664,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.013089599087834358,
+ "learning_rate": 0.001744671763136191,
+ "loss": 4.1626458168029785,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.01327646616846323,
+ "learning_rate": 0.0017433547763178665,
+ "loss": 4.189356803894043,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.014539042487740517,
+ "learning_rate": 0.0017420376055182233,
+ "loss": 4.176387786865234,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.013269098475575447,
+ "learning_rate": 0.001740720251798482,
+ "loss": 4.180716514587402,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.010922391898930073,
+ "learning_rate": 0.0017394027162200105,
+ "loss": 4.163928031921387,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.013355271890759468,
+ "learning_rate": 0.0017380849998443243,
+ "loss": 4.150012016296387,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.014149201102554798,
+ "learning_rate": 0.0017367671037330828,
+ "loss": 4.173713207244873,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01354879979044199,
+ "learning_rate": 0.0017354490289480918,
+ "loss": 4.150563716888428,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.013788704760372639,
+ "learning_rate": 0.0017341307765512998,
+ "loss": 4.213858604431152,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.011597222648561,
+ "learning_rate": 0.0017328123476047994,
+ "loss": 4.166989326477051,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.01055675558745861,
+ "learning_rate": 0.001731493743170824,
+ "loss": 4.172226428985596,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.011600397527217865,
+ "learning_rate": 0.0017301749643117506,
+ "loss": 4.1656951904296875,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.01114651933312416,
+ "learning_rate": 0.0017288560120900943,
+ "loss": 4.189825057983398,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.009811583906412125,
+ "learning_rate": 0.0017275368875685126,
+ "loss": 4.181164741516113,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.009758968837559223,
+ "learning_rate": 0.0017262175918097992,
+ "loss": 4.163188934326172,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.00999707542359829,
+ "learning_rate": 0.001724898125876887,
+ "loss": 4.184117317199707,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.009528180584311485,
+ "learning_rate": 0.001723578490832846,
+ "loss": 4.16407585144043,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.009717358276247978,
+ "learning_rate": 0.0017222586877408826,
+ "loss": 4.209892272949219,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.010909050703048706,
+ "learning_rate": 0.0017209387176643377,
+ "loss": 4.19977331161499,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.011234750039875507,
+ "learning_rate": 0.0017196185816666875,
+ "loss": 4.19363260269165,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.010300721041858196,
+ "learning_rate": 0.0017182982808115425,
+ "loss": 4.142427921295166,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.00941646471619606,
+ "learning_rate": 0.001716977816162644,
+ "loss": 4.159372329711914,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.009304054081439972,
+ "learning_rate": 0.0017156571887838679,
+ "loss": 4.207922458648682,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.009496180340647697,
+ "learning_rate": 0.0017143363997392184,
+ "loss": 4.137894630432129,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.009257436729967594,
+ "learning_rate": 0.0017130154500928326,
+ "loss": 4.198753833770752,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.00900586973875761,
+ "learning_rate": 0.0017116943409089746,
+ "loss": 4.160684585571289,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.009398494847118855,
+ "learning_rate": 0.00171037307325204,
+ "loss": 4.1819353103637695,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.008633013814687729,
+ "learning_rate": 0.0017090516481865483,
+ "loss": 4.176833152770996,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.00898425467312336,
+ "learning_rate": 0.0017077300667771484,
+ "loss": 4.190657138824463,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.008182398043572903,
+ "learning_rate": 0.0017064083300886155,
+ "loss": 4.173429489135742,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.009589143097400665,
+ "learning_rate": 0.0017050864391858478,
+ "loss": 4.1559014320373535,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.010304383002221584,
+ "learning_rate": 0.0017037643951338698,
+ "loss": 4.155453681945801,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.009943748824298382,
+ "learning_rate": 0.0017024421989978275,
+ "loss": 4.119734287261963,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.009238033555448055,
+ "learning_rate": 0.001701119851842992,
+ "loss": 4.188974857330322,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.008058690465986729,
+ "learning_rate": 0.0016997973547347531,
+ "loss": 4.19934606552124,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.008485953323543072,
+ "learning_rate": 0.0016984747087386233,
+ "loss": 4.165615558624268,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.008476605638861656,
+ "learning_rate": 0.0016971519149202347,
+ "loss": 4.21021842956543,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.008588708005845547,
+ "learning_rate": 0.0016958289743453385,
+ "loss": 4.208064556121826,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.009223178029060364,
+ "learning_rate": 0.0016945058880798037,
+ "loss": 4.158413887023926,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.009196167811751366,
+ "learning_rate": 0.0016931826571896166,
+ "loss": 4.203187465667725,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.009488921612501144,
+ "learning_rate": 0.001691859282740882,
+ "loss": 4.178227424621582,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.009253161028027534,
+ "learning_rate": 0.0016905357657998167,
+ "loss": 4.131230354309082,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.010069048963487148,
+ "learning_rate": 0.0016892121074327559,
+ "loss": 4.210734844207764,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.009945095516741276,
+ "learning_rate": 0.001687888308706147,
+ "loss": 4.184042930603027,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.010483640246093273,
+ "learning_rate": 0.0016865643706865505,
+ "loss": 4.156919479370117,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.011560715734958649,
+ "learning_rate": 0.0016852402944406391,
+ "loss": 4.185520172119141,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.010001277551054955,
+ "learning_rate": 0.0016839160810351978,
+ "loss": 4.226828575134277,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.010335934348404408,
+ "learning_rate": 0.0016825917315371202,
+ "loss": 4.183531284332275,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.011322776786983013,
+ "learning_rate": 0.001681267247013412,
+ "loss": 4.179635047912598,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.009663194417953491,
+ "learning_rate": 0.001679942628531186,
+ "loss": 4.165654182434082,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.010590246878564358,
+ "learning_rate": 0.0016786178771576643,
+ "loss": 4.1539459228515625,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.511386394500732,
+ "eval_runtime": 51.7751,
+ "eval_samples_per_second": 47.166,
+ "eval_steps_per_second": 1.487,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.01254354789853096,
+ "learning_rate": 0.0016772929939601728,
+ "loss": 4.104002475738525,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.01300987508147955,
+ "learning_rate": 0.0016759679800061483,
+ "loss": 4.127395153045654,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.013325816951692104,
+ "learning_rate": 0.0016746428363631299,
+ "loss": 4.131229400634766,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.014302870258688927,
+ "learning_rate": 0.001673317564098761,
+ "loss": 4.079358100891113,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.013776068575680256,
+ "learning_rate": 0.0016719921642807905,
+ "loss": 4.132065773010254,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.010788560844957829,
+ "learning_rate": 0.0016706666379770689,
+ "loss": 4.134626388549805,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.009766630828380585,
+ "learning_rate": 0.0016693409862555493,
+ "loss": 4.1645426750183105,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.011760029010474682,
+ "learning_rate": 0.0016680152101842842,
+ "loss": 4.110664367675781,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.011047836393117905,
+ "learning_rate": 0.0016666893108314288,
+ "loss": 4.129222869873047,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.009989139623939991,
+ "learning_rate": 0.0016653632892652344,
+ "loss": 4.076835632324219,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.010033783502876759,
+ "learning_rate": 0.001664037146554054,
+ "loss": 4.11555290222168,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.00996427796781063,
+ "learning_rate": 0.001662710883766337,
+ "loss": 4.131463527679443,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.01044023409485817,
+ "learning_rate": 0.0016613845019706288,
+ "loss": 4.098750114440918,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.010033872909843922,
+ "learning_rate": 0.0016600580022355712,
+ "loss": 4.161447048187256,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.009329567663371563,
+ "learning_rate": 0.0016587313856299012,
+ "loss": 4.146491527557373,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.010061589069664478,
+ "learning_rate": 0.0016574046532224504,
+ "loss": 4.107517719268799,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.009491910226643085,
+ "learning_rate": 0.0016560778060821424,
+ "loss": 4.098226547241211,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.008529345504939556,
+ "learning_rate": 0.0016547508452779943,
+ "loss": 4.1247711181640625,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.00936342403292656,
+ "learning_rate": 0.0016534237718791136,
+ "loss": 4.120573043823242,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.010255641303956509,
+ "learning_rate": 0.0016520965869547008,
+ "loss": 4.089067459106445,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.011638758704066277,
+ "learning_rate": 0.001650769291574044,
+ "loss": 4.074847221374512,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.014707857742905617,
+ "learning_rate": 0.0016494418868065214,
+ "loss": 4.153377532958984,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.01650111749768257,
+ "learning_rate": 0.0016481143737215986,
+ "loss": 4.134062767028809,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.014833372086286545,
+ "learning_rate": 0.00164678675338883,
+ "loss": 4.1344828605651855,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.012257886119186878,
+ "learning_rate": 0.0016454590268778538,
+ "loss": 4.098813533782959,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.013508322648704052,
+ "learning_rate": 0.0016441311952583959,
+ "loss": 4.106297492980957,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.01256900280714035,
+ "learning_rate": 0.001642803259600267,
+ "loss": 4.143631935119629,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.009885406121611595,
+ "learning_rate": 0.00164147522097336,
+ "loss": 4.120885848999023,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.012004674412310123,
+ "learning_rate": 0.0016401470804476518,
+ "loss": 4.089243412017822,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01224028691649437,
+ "learning_rate": 0.0016388188390932012,
+ "loss": 4.077765464782715,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.012511290609836578,
+ "learning_rate": 0.001637490497980149,
+ "loss": 4.161346435546875,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.010701716877520084,
+ "learning_rate": 0.0016361620581787143,
+ "loss": 4.10698938369751,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.010713113471865654,
+ "learning_rate": 0.001634833520759198,
+ "loss": 4.114283561706543,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.011018907651305199,
+ "learning_rate": 0.001633504886791978,
+ "loss": 4.134284019470215,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.011279584839940071,
+ "learning_rate": 0.0016321761573475117,
+ "loss": 4.1456804275512695,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.010642931796610355,
+ "learning_rate": 0.001630847333496332,
+ "loss": 4.160918235778809,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.011069678701460361,
+ "learning_rate": 0.0016295184163090482,
+ "loss": 4.136838912963867,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.011007166467607021,
+ "learning_rate": 0.0016281894068563443,
+ "loss": 4.129153728485107,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.010055703110992908,
+ "learning_rate": 0.0016268603062089802,
+ "loss": 4.138665676116943,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.009839230217039585,
+ "learning_rate": 0.0016255311154377875,
+ "loss": 4.1366496086120605,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.011191432364284992,
+ "learning_rate": 0.0016242018356136714,
+ "loss": 4.129880905151367,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.012130341492593288,
+ "learning_rate": 0.0016228724678076095,
+ "loss": 4.1425981521606445,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.012707795947790146,
+ "learning_rate": 0.0016215430130906482,
+ "loss": 4.115452766418457,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.012707809917628765,
+ "learning_rate": 0.001620213472533906,
+ "loss": 4.122758865356445,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.01182397548109293,
+ "learning_rate": 0.001618883847208569,
+ "loss": 4.140275955200195,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.010865340940654278,
+ "learning_rate": 0.001617554138185893,
+ "loss": 4.137759208679199,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.010973592288792133,
+ "learning_rate": 0.0016162243465371998,
+ "loss": 4.138947010040283,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.011247226968407631,
+ "learning_rate": 0.001614894473333879,
+ "loss": 4.115516662597656,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.010837487876415253,
+ "learning_rate": 0.0016135645196473853,
+ "loss": 4.160427093505859,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.009096396155655384,
+ "learning_rate": 0.0016122344865492385,
+ "loss": 4.1200032234191895,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.010671247728168964,
+ "learning_rate": 0.0016109043751110223,
+ "loss": 4.097919464111328,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.0107562355697155,
+ "learning_rate": 0.0016095741864043831,
+ "loss": 4.096290588378906,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.008909327909350395,
+ "learning_rate": 0.0016082439215010305,
+ "loss": 4.171017646789551,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.010305124334990978,
+ "learning_rate": 0.0016069135814727343,
+ "loss": 4.1450276374816895,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.011235613375902176,
+ "learning_rate": 0.0016055831673913255,
+ "loss": 4.134704113006592,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.00825632456690073,
+ "learning_rate": 0.0016042526803286949,
+ "loss": 4.107443332672119,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.009056378155946732,
+ "learning_rate": 0.0016029221213567921,
+ "loss": 4.132656097412109,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.008835164830088615,
+ "learning_rate": 0.001601591491547624,
+ "loss": 4.124729633331299,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.00915333442389965,
+ "learning_rate": 0.0016002607919732557,
+ "loss": 4.119500160217285,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.008532539941370487,
+ "learning_rate": 0.0015989300237058066,
+ "loss": 4.141661643981934,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.009952311404049397,
+ "learning_rate": 0.0015975991878174533,
+ "loss": 4.140929222106934,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.008920297026634216,
+ "learning_rate": 0.0015962682853804268,
+ "loss": 4.155877113342285,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.008770049549639225,
+ "learning_rate": 0.0015949373174670104,
+ "loss": 4.126394271850586,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.0078084105625748634,
+ "learning_rate": 0.0015936062851495411,
+ "loss": 4.128876209259033,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.00816641841083765,
+ "learning_rate": 0.0015922751895004078,
+ "loss": 4.137410640716553,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.008895459584891796,
+ "learning_rate": 0.0015909440315920505,
+ "loss": 4.07957649230957,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.008551528677344322,
+ "learning_rate": 0.001589612812496958,
+ "loss": 4.104213714599609,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.008325488306581974,
+ "learning_rate": 0.0015882815332876707,
+ "loss": 4.1584792137146,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.009341313503682613,
+ "learning_rate": 0.0015869501950367744,
+ "loss": 4.1417436599731445,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.00999050959944725,
+ "learning_rate": 0.0015856187988169067,
+ "loss": 4.136049270629883,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.008889932185411453,
+ "learning_rate": 0.0015842873457007484,
+ "loss": 4.16257381439209,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.009109573438763618,
+ "learning_rate": 0.0015829558367610268,
+ "loss": 4.171441555023193,
+ "step": 1800
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 7.67403340333056e+17,
+ "train_batch_size": 16,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/baseline/checkpoint-1800/training_args.bin b/runs/baseline/checkpoint-1800/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7
--- /dev/null
+++ b/runs/baseline/checkpoint-1800/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97
+size 5265
diff --git a/runs/baseline/checkpoint-2160/chat_template.jinja b/runs/baseline/checkpoint-2160/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/baseline/checkpoint-2160/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/baseline/checkpoint-2160/config.json b/runs/baseline/checkpoint-2160/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/baseline/checkpoint-2160/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/baseline/checkpoint-2160/generation_config.json b/runs/baseline/checkpoint-2160/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/baseline/checkpoint-2160/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/baseline/checkpoint-2160/model.safetensors b/runs/baseline/checkpoint-2160/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..d6a9e9ce06875771d4d27d5d86b491e2f87bfe5a
--- /dev/null
+++ b/runs/baseline/checkpoint-2160/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b80a8ab78d106d60f21de0abe3f37e7784d6094041c106a808255e0e47720f82
+size 583356232
diff --git a/runs/baseline/checkpoint-2160/optimizer.pt b/runs/baseline/checkpoint-2160/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..8cfc40140a1ce500459fdb375065966569d80632
--- /dev/null
+++ b/runs/baseline/checkpoint-2160/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f45e69b0c66324b605ce8625895e3bf9f80169892b256fb0c7418e13fee28c4d
+size 1166825803
diff --git a/runs/baseline/checkpoint-2160/rng_state_0.pth b/runs/baseline/checkpoint-2160/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..86ca1aad69953afb19f9f8da4c01764c2f44d22b
--- /dev/null
+++ b/runs/baseline/checkpoint-2160/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:253e65ee29b5e996242fd0eef4ab59e40fe385a114b29c442de150cf74084b60
+size 14917
diff --git a/runs/baseline/checkpoint-2160/rng_state_1.pth b/runs/baseline/checkpoint-2160/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3031af55ab1520df411513147e482ece04c4d4f6
--- /dev/null
+++ b/runs/baseline/checkpoint-2160/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:04c9d382d23a06579a44352e657618aaf7b647b0f5de65751c81300c8f96f42a
+size 14917
diff --git a/runs/baseline/checkpoint-2160/scheduler.pt b/runs/baseline/checkpoint-2160/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..95e6d3366fe6af5dc063f7b8974daf6c734b2ea8
--- /dev/null
+++ b/runs/baseline/checkpoint-2160/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:63d15070a4477d2431b35b7fbfcc0f72990026ae265966082eb9b68e8bb7618c
+size 1465
diff --git a/runs/baseline/checkpoint-2160/tokenizer.json b/runs/baseline/checkpoint-2160/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/baseline/checkpoint-2160/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/baseline/checkpoint-2160/tokenizer_config.json b/runs/baseline/checkpoint-2160/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/baseline/checkpoint-2160/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/baseline/checkpoint-2160/trainer_state.json b/runs/baseline/checkpoint-2160/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..ff194229364d154928e8b2e342c3a7a06a6e42ef
--- /dev/null
+++ b/runs/baseline/checkpoint-2160/trainer_state.json
@@ -0,0 +1,15386 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 30.0,
+ "eval_steps": 500,
+ "global_step": 2160,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.12165174633264542,
+ "learning_rate": 0.0,
+ "loss": 12.027063369750977,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.12109383195638657,
+ "learning_rate": 3e-05,
+ "loss": 12.028635025024414,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.12690064311027527,
+ "learning_rate": 6e-05,
+ "loss": 11.894524574279785,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.1537286937236786,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.672262191772461,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.11778244376182556,
+ "learning_rate": 0.00012,
+ "loss": 11.433671951293945,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.10874935984611511,
+ "learning_rate": 0.00015000000000000001,
+ "loss": 11.270309448242188,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.10641921311616898,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 11.155481338500977,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.1048506498336792,
+ "learning_rate": 0.00021000000000000004,
+ "loss": 11.066675186157227,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.10546834021806717,
+ "learning_rate": 0.00024,
+ "loss": 10.952295303344727,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.10614268481731415,
+ "learning_rate": 0.00027,
+ "loss": 10.811386108398438,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10529010742902756,
+ "learning_rate": 0.00030000000000000003,
+ "loss": 10.657712936401367,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.1051444336771965,
+ "learning_rate": 0.00033,
+ "loss": 10.47591781616211,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10420846194028854,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 10.280141830444336,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10323625057935715,
+ "learning_rate": 0.00039000000000000005,
+ "loss": 10.076366424560547,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10255401581525803,
+ "learning_rate": 0.00042000000000000007,
+ "loss": 9.856939315795898,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10151027888059616,
+ "learning_rate": 0.00045,
+ "loss": 9.632134437561035,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.1006326898932457,
+ "learning_rate": 0.00048,
+ "loss": 9.401637077331543,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.09820562601089478,
+ "learning_rate": 0.00051,
+ "loss": 9.177400588989258,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.0956626906991005,
+ "learning_rate": 0.00054,
+ "loss": 8.95518684387207,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.0922091156244278,
+ "learning_rate": 0.00057,
+ "loss": 8.75066089630127,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.08593932539224625,
+ "learning_rate": 0.0006000000000000001,
+ "loss": 8.557976722717285,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.08014766871929169,
+ "learning_rate": 0.00063,
+ "loss": 8.360604286193848,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.07381244748830795,
+ "learning_rate": 0.00066,
+ "loss": 8.210375785827637,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.06455052644014359,
+ "learning_rate": 0.0006900000000000001,
+ "loss": 8.07168960571289,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.054115474224090576,
+ "learning_rate": 0.0007199999999999999,
+ "loss": 7.960069179534912,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.043575264513492584,
+ "learning_rate": 0.00075,
+ "loss": 7.850685119628906,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.02976679615676403,
+ "learning_rate": 0.0007800000000000001,
+ "loss": 7.824377059936523,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.020552165806293488,
+ "learning_rate": 0.0008100000000000001,
+ "loss": 7.7695417404174805,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.018530743196606636,
+ "learning_rate": 0.0008400000000000001,
+ "loss": 7.782247543334961,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.02716820128262043,
+ "learning_rate": 0.00087,
+ "loss": 7.754945278167725,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.03638409823179245,
+ "learning_rate": 0.0009,
+ "loss": 7.799381256103516,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.041579652577638626,
+ "learning_rate": 0.00093,
+ "loss": 7.776629447937012,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.0304960235953331,
+ "learning_rate": 0.00096,
+ "loss": 7.749439239501953,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.053728822618722916,
+ "learning_rate": 0.00099,
+ "loss": 7.745059490203857,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.036322493106126785,
+ "learning_rate": 0.00102,
+ "loss": 7.752074718475342,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.03232959285378456,
+ "learning_rate": 0.00105,
+ "loss": 7.704257011413574,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.03246184438467026,
+ "learning_rate": 0.00108,
+ "loss": 7.667045593261719,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.036655545234680176,
+ "learning_rate": 0.00111,
+ "loss": 7.623445987701416,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.021042725071310997,
+ "learning_rate": 0.00114,
+ "loss": 7.582821846008301,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.022982975468039513,
+ "learning_rate": 0.00117,
+ "loss": 7.58036994934082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.024781864136457443,
+ "learning_rate": 0.0012000000000000001,
+ "loss": 7.5903000831604,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.03667089343070984,
+ "learning_rate": 0.00123,
+ "loss": 7.588233947753906,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.04083017632365227,
+ "learning_rate": 0.00126,
+ "loss": 7.548343658447266,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.02585385926067829,
+ "learning_rate": 0.00129,
+ "loss": 7.604833602905273,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.018793931230902672,
+ "learning_rate": 0.00132,
+ "loss": 7.56705379486084,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.025538936257362366,
+ "learning_rate": 0.00135,
+ "loss": 7.562107086181641,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.02819318138062954,
+ "learning_rate": 0.0013800000000000002,
+ "loss": 7.517213821411133,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.016239255666732788,
+ "learning_rate": 0.00141,
+ "loss": 7.4780755043029785,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.024470701813697815,
+ "learning_rate": 0.0014399999999999999,
+ "loss": 7.505887985229492,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.015499847941100597,
+ "learning_rate": 0.00147,
+ "loss": 7.484469890594482,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.017785200849175453,
+ "learning_rate": 0.0015,
+ "loss": 7.481822967529297,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.015952805057168007,
+ "learning_rate": 0.0015300000000000001,
+ "loss": 7.484804153442383,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.015771260485053062,
+ "learning_rate": 0.0015600000000000002,
+ "loss": 7.4495439529418945,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.024370530620217323,
+ "learning_rate": 0.00159,
+ "loss": 7.471208572387695,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.018173737451434135,
+ "learning_rate": 0.0016200000000000001,
+ "loss": 7.438447952270508,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.020564094185829163,
+ "learning_rate": 0.0016500000000000002,
+ "loss": 7.437922954559326,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.016154944896697998,
+ "learning_rate": 0.0016800000000000003,
+ "loss": 7.429378032684326,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.016817007213830948,
+ "learning_rate": 0.00171,
+ "loss": 7.406428337097168,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.03460947796702385,
+ "learning_rate": 0.00174,
+ "loss": 7.393331527709961,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.15410488843917847,
+ "learning_rate": 0.0017699999999999999,
+ "loss": 7.645856857299805,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.07443148642778397,
+ "learning_rate": 0.0018,
+ "loss": 7.549508094787598,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.0507572740316391,
+ "learning_rate": 0.00183,
+ "loss": 7.517921447753906,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.09682746231555939,
+ "learning_rate": 0.00186,
+ "loss": 7.543978691101074,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.2571517527103424,
+ "learning_rate": 0.00189,
+ "loss": 7.768949508666992,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.0317784883081913,
+ "learning_rate": 0.00192,
+ "loss": 7.509063720703125,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.0353970006108284,
+ "learning_rate": 0.0019500000000000001,
+ "loss": 7.522322654724121,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.03663994371891022,
+ "learning_rate": 0.00198,
+ "loss": 7.504723072052002,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.025481393560767174,
+ "learning_rate": 0.00201,
+ "loss": 7.452597141265869,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.03099915198981762,
+ "learning_rate": 0.00204,
+ "loss": 7.455441474914551,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.022921698167920113,
+ "learning_rate": 0.00207,
+ "loss": 7.4454827308654785,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.032388731837272644,
+ "learning_rate": 0.0021,
+ "loss": 7.467103958129883,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.022428609430789948,
+ "learning_rate": 0.00213,
+ "loss": 7.467540264129639,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.469719409942627,
+ "eval_runtime": 50.9714,
+ "eval_samples_per_second": 47.909,
+ "eval_steps_per_second": 1.511,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.017480911687016487,
+ "learning_rate": 0.00216,
+ "loss": 7.455810546875,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.016675131395459175,
+ "learning_rate": 0.00219,
+ "loss": 7.474848747253418,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.017450228333473206,
+ "learning_rate": 0.00222,
+ "loss": 7.442611217498779,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.016823675483465195,
+ "learning_rate": 0.0022500000000000003,
+ "loss": 7.454628944396973,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.016717543825507164,
+ "learning_rate": 0.00228,
+ "loss": 7.425317764282227,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.01644429750740528,
+ "learning_rate": 0.00231,
+ "loss": 7.4102864265441895,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.013104400597512722,
+ "learning_rate": 0.00234,
+ "loss": 7.4559173583984375,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.01175684668123722,
+ "learning_rate": 0.00237,
+ "loss": 7.438898086547852,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.013736684806644917,
+ "learning_rate": 0.0024000000000000002,
+ "loss": 7.449005603790283,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01176756713539362,
+ "learning_rate": 0.0024300000000000003,
+ "loss": 7.428421974182129,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.011860918253660202,
+ "learning_rate": 0.00246,
+ "loss": 7.411491394042969,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.010093477554619312,
+ "learning_rate": 0.00249,
+ "loss": 7.39311408996582,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.009453288279473782,
+ "learning_rate": 0.00252,
+ "loss": 7.387656211853027,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.010311692021787167,
+ "learning_rate": 0.00255,
+ "loss": 7.401076316833496,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.010053195990622044,
+ "learning_rate": 0.00258,
+ "loss": 7.39655876159668,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.009758037514984608,
+ "learning_rate": 0.00261,
+ "loss": 7.423122882843018,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008394946344196796,
+ "learning_rate": 0.00264,
+ "loss": 7.383601188659668,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.008457187563180923,
+ "learning_rate": 0.00267,
+ "loss": 7.400934219360352,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.010601334273815155,
+ "learning_rate": 0.0027,
+ "loss": 7.400241851806641,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.009009373374283314,
+ "learning_rate": 0.0027300000000000002,
+ "loss": 7.377348899841309,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.00737508200109005,
+ "learning_rate": 0.0027600000000000003,
+ "loss": 7.385881423950195,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.0075711836107075214,
+ "learning_rate": 0.0027900000000000004,
+ "loss": 7.348570823669434,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.00846901722252369,
+ "learning_rate": 0.00282,
+ "loss": 7.364287853240967,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.007160635199397802,
+ "learning_rate": 0.00285,
+ "loss": 7.3881635665893555,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.00682067358866334,
+ "learning_rate": 0.0028799999999999997,
+ "loss": 7.348663330078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.009108101949095726,
+ "learning_rate": 0.00291,
+ "loss": 7.356302261352539,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.015828927978873253,
+ "learning_rate": 0.00294,
+ "loss": 7.347576141357422,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.009683096781373024,
+ "learning_rate": 0.00297,
+ "loss": 7.323967933654785,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.010833270847797394,
+ "learning_rate": 0.003,
+ "loss": 7.3139495849609375,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.014905400574207306,
+ "learning_rate": 0.002999999401781161,
+ "loss": 7.302030086517334,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.023674853146076202,
+ "learning_rate": 0.002999997607125127,
+ "loss": 7.311811923980713,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.050460800528526306,
+ "learning_rate": 0.002999994616033342,
+ "loss": 7.41972541809082,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.02193174883723259,
+ "learning_rate": 0.002999990428508218,
+ "loss": 7.294468879699707,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.036261629313230515,
+ "learning_rate": 0.0029999850445531283,
+ "loss": 7.295568466186523,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.02734929881989956,
+ "learning_rate": 0.002999978464172409,
+ "loss": 7.309239864349365,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.017884107306599617,
+ "learning_rate": 0.002999970687371363,
+ "loss": 7.2427592277526855,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01700308918952942,
+ "learning_rate": 0.0029999617141562573,
+ "loss": 7.241730213165283,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.0209935512393713,
+ "learning_rate": 0.0029999515445343194,
+ "loss": 7.262085437774658,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.023207105696201324,
+ "learning_rate": 0.0029999401785137437,
+ "loss": 7.25450325012207,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.01907353848218918,
+ "learning_rate": 0.002999927616103687,
+ "loss": 7.23149299621582,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.016149666160345078,
+ "learning_rate": 0.0029999138573142715,
+ "loss": 7.210232734680176,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.017513928934931755,
+ "learning_rate": 0.002999898902156582,
+ "loss": 7.218451976776123,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.015002547763288021,
+ "learning_rate": 0.002999882750642667,
+ "loss": 7.169316291809082,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.014233081601560116,
+ "learning_rate": 0.0029998654027855406,
+ "loss": 7.194825172424316,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.01283260714262724,
+ "learning_rate": 0.0029998468585991783,
+ "loss": 7.15669059753418,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.014416799880564213,
+ "learning_rate": 0.0029998271180985223,
+ "loss": 7.143369674682617,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.014337058179080486,
+ "learning_rate": 0.0029998061812994757,
+ "loss": 7.127734184265137,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.016406210139393806,
+ "learning_rate": 0.002999784048218908,
+ "loss": 7.086504936218262,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.01680421642959118,
+ "learning_rate": 0.002999760718874652,
+ "loss": 7.089420318603516,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.010665548034012318,
+ "learning_rate": 0.0029997361932855016,
+ "loss": 7.102463722229004,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013479137793183327,
+ "learning_rate": 0.002999710471471218,
+ "loss": 7.079098701477051,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.012408486567437649,
+ "learning_rate": 0.0029996835534525246,
+ "loss": 7.0790114402771,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.012035196647047997,
+ "learning_rate": 0.0029996554392511094,
+ "loss": 7.058756351470947,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.00899985246360302,
+ "learning_rate": 0.0029996261288896225,
+ "loss": 7.0739240646362305,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.01072906143963337,
+ "learning_rate": 0.002999595622391679,
+ "loss": 7.085537910461426,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.01047342736274004,
+ "learning_rate": 0.0029995639197818577,
+ "loss": 7.060049057006836,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.02342618815600872,
+ "learning_rate": 0.0029995310210857006,
+ "loss": 7.039482116699219,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.011256859637796879,
+ "learning_rate": 0.0029994969263297134,
+ "loss": 7.040412902832031,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.008407019078731537,
+ "learning_rate": 0.0029994616355413666,
+ "loss": 7.0364274978637695,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.011006148532032967,
+ "learning_rate": 0.0029994251487490916,
+ "loss": 6.991468906402588,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.009103255346417427,
+ "learning_rate": 0.0029993874659822875,
+ "loss": 7.001059532165527,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.007875788025557995,
+ "learning_rate": 0.0029993485872713125,
+ "loss": 6.9974260330200195,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.007872611284255981,
+ "learning_rate": 0.0029993085126474913,
+ "loss": 6.980216979980469,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.010071957483887672,
+ "learning_rate": 0.0029992672421431115,
+ "loss": 6.97799015045166,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.01209618803113699,
+ "learning_rate": 0.0029992247757914237,
+ "loss": 6.97202205657959,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.014884858392179012,
+ "learning_rate": 0.0029991811136266425,
+ "loss": 6.976570129394531,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.014953463338315487,
+ "learning_rate": 0.0029991362556839453,
+ "loss": 6.971050262451172,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.008371063508093357,
+ "learning_rate": 0.002999090201999474,
+ "loss": 6.955556869506836,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.011926881037652493,
+ "learning_rate": 0.002999042952610332,
+ "loss": 6.943947792053223,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.013422893360257149,
+ "learning_rate": 0.002998994507554589,
+ "loss": 6.910022735595703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.009672224521636963,
+ "learning_rate": 0.002998944866871275,
+ "loss": 6.8994855880737305,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.012590581551194191,
+ "learning_rate": 0.0029988940306003853,
+ "loss": 6.88709020614624,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.945146083831787,
+ "eval_runtime": 51.5801,
+ "eval_samples_per_second": 47.344,
+ "eval_steps_per_second": 1.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.016554836183786392,
+ "learning_rate": 0.002998841998782877,
+ "loss": 6.916609764099121,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.011085660196840763,
+ "learning_rate": 0.0029987887714606723,
+ "loss": 6.884910583496094,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.008596168830990791,
+ "learning_rate": 0.0029987343486766547,
+ "loss": 6.900066375732422,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.014487302862107754,
+ "learning_rate": 0.0029986787304746715,
+ "loss": 6.8693342208862305,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.015636997297406197,
+ "learning_rate": 0.002998621916899534,
+ "loss": 6.903670787811279,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.013275462202727795,
+ "learning_rate": 0.0029985639079970153,
+ "loss": 6.889361381530762,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.009044744074344635,
+ "learning_rate": 0.002998504703813852,
+ "loss": 6.8549981117248535,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.011327696964144707,
+ "learning_rate": 0.002998444304397745,
+ "loss": 6.856151103973389,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.01310331467539072,
+ "learning_rate": 0.0029983827097973554,
+ "loss": 6.834944725036621,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.010003753937780857,
+ "learning_rate": 0.00299831992006231,
+ "loss": 6.825613975524902,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.009928115643560886,
+ "learning_rate": 0.0029982559352431967,
+ "loss": 6.828935146331787,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.012302150949835777,
+ "learning_rate": 0.0029981907553915674,
+ "loss": 6.821207046508789,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.020555168390274048,
+ "learning_rate": 0.002998124380559936,
+ "loss": 6.8184895515441895,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.011417638510465622,
+ "learning_rate": 0.0029980568108017797,
+ "loss": 6.788754463195801,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.010772978886961937,
+ "learning_rate": 0.002997988046171538,
+ "loss": 6.7861433029174805,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.012753859162330627,
+ "learning_rate": 0.002997918086724614,
+ "loss": 6.788776397705078,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.011327114887535572,
+ "learning_rate": 0.0029978469325173717,
+ "loss": 6.779160022735596,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01235753484070301,
+ "learning_rate": 0.002997774583607139,
+ "loss": 6.799554824829102,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.01417156495153904,
+ "learning_rate": 0.0029977010400522074,
+ "loss": 6.77957820892334,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.021321553736925125,
+ "learning_rate": 0.002997626301911828,
+ "loss": 6.785632133483887,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.03614472970366478,
+ "learning_rate": 0.002997550369246217,
+ "loss": 6.836414337158203,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03934052586555481,
+ "learning_rate": 0.002997473242116551,
+ "loss": 6.854659557342529,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.016240060329437256,
+ "learning_rate": 0.0029973949205849707,
+ "loss": 6.789690971374512,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.031164905056357384,
+ "learning_rate": 0.0029973154047145782,
+ "loss": 6.8363776206970215,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.013051844201982021,
+ "learning_rate": 0.002997234694569438,
+ "loss": 6.803576469421387,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.022611448541283607,
+ "learning_rate": 0.002997152790214577,
+ "loss": 6.762107849121094,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.017812907695770264,
+ "learning_rate": 0.0029970696917159833,
+ "loss": 6.776218414306641,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.015453332103788853,
+ "learning_rate": 0.0029969853991406088,
+ "loss": 6.742618083953857,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.020630091428756714,
+ "learning_rate": 0.0029968999125563655,
+ "loss": 6.765475749969482,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.014672341756522655,
+ "learning_rate": 0.0029968132320321297,
+ "loss": 6.754009246826172,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.016540220007300377,
+ "learning_rate": 0.0029967253576377376,
+ "loss": 6.734752655029297,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.013495649211108685,
+ "learning_rate": 0.0029966362894439874,
+ "loss": 6.752676010131836,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015936989337205887,
+ "learning_rate": 0.0029965460275226405,
+ "loss": 6.725708961486816,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010431243106722832,
+ "learning_rate": 0.0029964545719464194,
+ "loss": 6.698376655578613,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01669321395456791,
+ "learning_rate": 0.0029963619227890083,
+ "loss": 6.711770534515381,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.013308997265994549,
+ "learning_rate": 0.0029962680801250514,
+ "loss": 6.705286979675293,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011648476123809814,
+ "learning_rate": 0.002996173044030159,
+ "loss": 6.690052509307861,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.03942161798477173,
+ "learning_rate": 0.0029960768145808965,
+ "loss": 6.711885452270508,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.01571274921298027,
+ "learning_rate": 0.002995979391854797,
+ "loss": 6.715847969055176,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.00829714722931385,
+ "learning_rate": 0.0029958807759303515,
+ "loss": 6.685308933258057,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.01374727487564087,
+ "learning_rate": 0.0029957809668870114,
+ "loss": 6.695744514465332,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.010857059620320797,
+ "learning_rate": 0.0029956799648051934,
+ "loss": 6.658294677734375,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.008897246792912483,
+ "learning_rate": 0.002995577769766272,
+ "loss": 6.642439842224121,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015880772843956947,
+ "learning_rate": 0.002995474381852584,
+ "loss": 6.674907684326172,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.021623898297548294,
+ "learning_rate": 0.0029953698011474265,
+ "loss": 6.66236686706543,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.03084627166390419,
+ "learning_rate": 0.0029952640277350596,
+ "loss": 6.666715621948242,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.040561504662036896,
+ "learning_rate": 0.002995157061700702,
+ "loss": 6.649062156677246,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.031126515939831734,
+ "learning_rate": 0.0029950489031305345,
+ "loss": 6.640542030334473,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.00872360821813345,
+ "learning_rate": 0.0029949395521116987,
+ "loss": 6.635221481323242,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.023432321846485138,
+ "learning_rate": 0.0029948290087322966,
+ "loss": 6.642196178436279,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.021096013486385345,
+ "learning_rate": 0.0029947172730813913,
+ "loss": 6.643655300140381,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.014494858682155609,
+ "learning_rate": 0.002994604345249006,
+ "loss": 6.612720966339111,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.05225599184632301,
+ "learning_rate": 0.0029944902253261248,
+ "loss": 6.60659646987915,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.014311269856989384,
+ "learning_rate": 0.0029943749134046915,
+ "loss": 6.54373025894165,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.010777472518384457,
+ "learning_rate": 0.0029942584095776115,
+ "loss": 6.599837303161621,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.012498079799115658,
+ "learning_rate": 0.002994140713938749,
+ "loss": 6.589657783508301,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.015775132924318314,
+ "learning_rate": 0.00299402182658293,
+ "loss": 6.577511787414551,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.017978230491280556,
+ "learning_rate": 0.0029939017476059396,
+ "loss": 6.58717155456543,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.026807745918631554,
+ "learning_rate": 0.0029937804771045236,
+ "loss": 6.578145503997803,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.04853282868862152,
+ "learning_rate": 0.0029936580151763872,
+ "loss": 6.609755516052246,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.1057916209101677,
+ "learning_rate": 0.002993534361920195,
+ "loss": 6.870367050170898,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.04925911873579025,
+ "learning_rate": 0.002993409517435574,
+ "loss": 6.871556282043457,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.05288556218147278,
+ "learning_rate": 0.0029932834818231077,
+ "loss": 6.808315753936768,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.03831710293889046,
+ "learning_rate": 0.0029931562551843405,
+ "loss": 6.827617645263672,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.03665117174386978,
+ "learning_rate": 0.002993027837621778,
+ "loss": 6.767284870147705,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.026216736063361168,
+ "learning_rate": 0.0029928982292388826,
+ "loss": 6.733659744262695,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.024796854704618454,
+ "learning_rate": 0.002992767430140078,
+ "loss": 6.759577751159668,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.02410992607474327,
+ "learning_rate": 0.002992635440430747,
+ "loss": 6.701825141906738,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.019924134016036987,
+ "learning_rate": 0.0029925022602172308,
+ "loss": 6.677121162414551,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.022516867145895958,
+ "learning_rate": 0.00299236788960683,
+ "loss": 6.668622970581055,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.025068556889891624,
+ "learning_rate": 0.002992232328707806,
+ "loss": 6.67718505859375,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.02574208751320839,
+ "learning_rate": 0.002992095577629376,
+ "loss": 6.679594039916992,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.657555103302002,
+ "eval_runtime": 51.0199,
+ "eval_samples_per_second": 47.864,
+ "eval_steps_per_second": 1.509,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.017298126593232155,
+ "learning_rate": 0.0029919576364817184,
+ "loss": 6.620774269104004,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.018840325996279716,
+ "learning_rate": 0.002991818505375971,
+ "loss": 6.613316535949707,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.020813113078475,
+ "learning_rate": 0.002991678184424228,
+ "loss": 6.621555328369141,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.026581479236483574,
+ "learning_rate": 0.0029915366737395437,
+ "loss": 6.613471031188965,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.02084030583500862,
+ "learning_rate": 0.0029913939734359306,
+ "loss": 6.559704780578613,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.01327502354979515,
+ "learning_rate": 0.00299125008362836,
+ "loss": 6.564910888671875,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.015091734007000923,
+ "learning_rate": 0.0029911050044327614,
+ "loss": 6.571262836456299,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.015463982708752155,
+ "learning_rate": 0.002990958735966023,
+ "loss": 6.565750598907471,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.013636826537549496,
+ "learning_rate": 0.0029908112783459894,
+ "loss": 6.57981014251709,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.010613273829221725,
+ "learning_rate": 0.0029906626316914655,
+ "loss": 6.538505554199219,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.01804237812757492,
+ "learning_rate": 0.0029905127961222127,
+ "loss": 6.5282511711120605,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.015131509862840176,
+ "learning_rate": 0.0029903617717589506,
+ "loss": 6.531335830688477,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.010518206283450127,
+ "learning_rate": 0.002990209558723358,
+ "loss": 6.521886825561523,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.011360022239387035,
+ "learning_rate": 0.002990056157138069,
+ "loss": 6.528301239013672,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.011556181125342846,
+ "learning_rate": 0.0029899015671266773,
+ "loss": 6.497917652130127,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.01365791168063879,
+ "learning_rate": 0.0029897457888137325,
+ "loss": 6.52849006652832,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.01091043371707201,
+ "learning_rate": 0.0029895888223247426,
+ "loss": 6.492197036743164,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.008484847843647003,
+ "learning_rate": 0.002989430667786173,
+ "loss": 6.497377395629883,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.00836879201233387,
+ "learning_rate": 0.0029892713253254466,
+ "loss": 6.495189666748047,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.009781497530639172,
+ "learning_rate": 0.0029891107950709406,
+ "loss": 6.454085826873779,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.009603648446500301,
+ "learning_rate": 0.0029889490771519934,
+ "loss": 6.499475479125977,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.007991855964064598,
+ "learning_rate": 0.0029887861716988975,
+ "loss": 6.460785865783691,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.008447187952697277,
+ "learning_rate": 0.0029886220788429025,
+ "loss": 6.504582405090332,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.010945498012006283,
+ "learning_rate": 0.0029884567987162156,
+ "loss": 6.49589729309082,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.011096215806901455,
+ "learning_rate": 0.0029882903314519997,
+ "loss": 6.460657119750977,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.013094739057123661,
+ "learning_rate": 0.0029881226771843747,
+ "loss": 6.462450981140137,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.015423037111759186,
+ "learning_rate": 0.0029879538360484165,
+ "loss": 6.45565128326416,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.0182222630828619,
+ "learning_rate": 0.002987783808180157,
+ "loss": 6.452227592468262,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.01981227844953537,
+ "learning_rate": 0.002987612593716585,
+ "loss": 6.477701663970947,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.026733700186014175,
+ "learning_rate": 0.002987440192795645,
+ "loss": 6.452212333679199,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.03504155948758125,
+ "learning_rate": 0.0029872666055562372,
+ "loss": 6.509665489196777,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.027970125898718834,
+ "learning_rate": 0.002987091832138217,
+ "loss": 6.487089157104492,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.013714591972529888,
+ "learning_rate": 0.0029869158726823974,
+ "loss": 6.469569206237793,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.016472695395350456,
+ "learning_rate": 0.0029867387273305453,
+ "loss": 6.461688995361328,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.018457358703017235,
+ "learning_rate": 0.0029865603962253833,
+ "loss": 6.490985870361328,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.011045100167393684,
+ "learning_rate": 0.0029863808795105896,
+ "loss": 6.403701305389404,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.014126897789537907,
+ "learning_rate": 0.0029862001773307972,
+ "loss": 6.379804611206055,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.014015335589647293,
+ "learning_rate": 0.0029860182898315954,
+ "loss": 6.404377460479736,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.010032658465206623,
+ "learning_rate": 0.002985835217159527,
+ "loss": 6.40291690826416,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.015469838865101337,
+ "learning_rate": 0.0029856509594620912,
+ "loss": 6.4414520263671875,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.011106910184025764,
+ "learning_rate": 0.0029854655168877397,
+ "loss": 6.396444320678711,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.01262320950627327,
+ "learning_rate": 0.0029852788895858816,
+ "loss": 6.385222434997559,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.012783976271748543,
+ "learning_rate": 0.002985091077706878,
+ "loss": 6.3818135261535645,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.01224067248404026,
+ "learning_rate": 0.0029849020814020468,
+ "loss": 6.37886905670166,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.010540086776018143,
+ "learning_rate": 0.0029847119008236576,
+ "loss": 6.37799072265625,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.012479913420975208,
+ "learning_rate": 0.0029845205361249366,
+ "loss": 6.365094184875488,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.010080077685415745,
+ "learning_rate": 0.0029843279874600622,
+ "loss": 6.371852874755859,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.013266298919916153,
+ "learning_rate": 0.002984134254984168,
+ "loss": 6.380039215087891,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.01014771405607462,
+ "learning_rate": 0.002983939338853339,
+ "loss": 6.370337963104248,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.008164579048752785,
+ "learning_rate": 0.0029837432392246185,
+ "loss": 6.331899642944336,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.01020442508161068,
+ "learning_rate": 0.002983545956255998,
+ "loss": 6.35077428817749,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.009898579679429531,
+ "learning_rate": 0.002983347490106426,
+ "loss": 6.371985912322998,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.009668919257819653,
+ "learning_rate": 0.002983147840935802,
+ "loss": 6.356776237487793,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.012560366652905941,
+ "learning_rate": 0.0029829470089049815,
+ "loss": 6.357911109924316,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.018435249105095863,
+ "learning_rate": 0.0029827449941757698,
+ "loss": 6.366821765899658,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.029046855866909027,
+ "learning_rate": 0.002982541796910927,
+ "loss": 6.367595672607422,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.043010637164115906,
+ "learning_rate": 0.002982337417274165,
+ "loss": 6.413588523864746,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.045470818877220154,
+ "learning_rate": 0.0029821318554301497,
+ "loss": 6.435015678405762,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.02148726023733616,
+ "learning_rate": 0.002981925111544498,
+ "loss": 6.330894470214844,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.026808688417077065,
+ "learning_rate": 0.00298171718578378,
+ "loss": 6.384839057922363,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.018495213240385056,
+ "learning_rate": 0.002981508078315517,
+ "loss": 6.36277961730957,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.025393597781658173,
+ "learning_rate": 0.0029812977893081834,
+ "loss": 6.3702239990234375,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020856289193034172,
+ "learning_rate": 0.002981086318931206,
+ "loss": 6.354547500610352,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.017827816307544708,
+ "learning_rate": 0.002980873667354962,
+ "loss": 6.339264869689941,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.02639714442193508,
+ "learning_rate": 0.002980659834750781,
+ "loss": 6.3254194259643555,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.015238096937537193,
+ "learning_rate": 0.002980444821290944,
+ "loss": 6.365842819213867,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.012863729149103165,
+ "learning_rate": 0.002980228627148684,
+ "loss": 6.356776714324951,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.014739899896085262,
+ "learning_rate": 0.0029800112524981845,
+ "loss": 6.332481384277344,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.013582171872258186,
+ "learning_rate": 0.0029797926975145797,
+ "loss": 6.311968803405762,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.012704637832939625,
+ "learning_rate": 0.0029795729623739555,
+ "loss": 6.321654319763184,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.012550560757517815,
+ "learning_rate": 0.0029793520472533495,
+ "loss": 6.298409461975098,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.011170165613293648,
+ "learning_rate": 0.002979129952330748,
+ "loss": 6.294830799102783,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.326757907867432,
+ "eval_runtime": 51.6566,
+ "eval_samples_per_second": 47.274,
+ "eval_steps_per_second": 1.491,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.010947446338832378,
+ "learning_rate": 0.0029789066777850897,
+ "loss": 6.286581039428711,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.010097995400428772,
+ "learning_rate": 0.002978682223796262,
+ "loss": 6.288301467895508,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.01224475260823965,
+ "learning_rate": 0.002978456590545104,
+ "loss": 6.264986991882324,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.009734537452459335,
+ "learning_rate": 0.0029782297782134045,
+ "loss": 6.304243087768555,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.01136956550180912,
+ "learning_rate": 0.0029780017869839013,
+ "loss": 6.244938373565674,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.00935293547809124,
+ "learning_rate": 0.002977772617040284,
+ "loss": 6.251852035522461,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.011118784546852112,
+ "learning_rate": 0.0029775422685671894,
+ "loss": 6.256633758544922,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.010837923735380173,
+ "learning_rate": 0.0029773107417502055,
+ "loss": 6.263411045074463,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.011920305900275707,
+ "learning_rate": 0.0029770780367758703,
+ "loss": 6.239311218261719,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.014117483049631119,
+ "learning_rate": 0.0029768441538316685,
+ "loss": 6.253314018249512,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.022502148523926735,
+ "learning_rate": 0.002976609093106036,
+ "loss": 6.246156692504883,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.03348720073699951,
+ "learning_rate": 0.002976372854788357,
+ "loss": 6.255914688110352,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.03427121415734291,
+ "learning_rate": 0.002976135439068965,
+ "loss": 6.289946556091309,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.02407355234026909,
+ "learning_rate": 0.002975896846139141,
+ "loss": 6.288534164428711,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.017861567437648773,
+ "learning_rate": 0.0029756570761911147,
+ "loss": 6.258049964904785,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.019980227574706078,
+ "learning_rate": 0.0029754161294180646,
+ "loss": 6.250895023345947,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.01443127728998661,
+ "learning_rate": 0.0029751740060141177,
+ "loss": 6.212200164794922,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.014975055120885372,
+ "learning_rate": 0.0029749307061743475,
+ "loss": 6.230568885803223,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01408296450972557,
+ "learning_rate": 0.0029746862300947773,
+ "loss": 6.243950366973877,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.011649327352643013,
+ "learning_rate": 0.0029744405779723766,
+ "loss": 6.197744846343994,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.013392188586294651,
+ "learning_rate": 0.002974193750005063,
+ "loss": 6.207277297973633,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.010902997106313705,
+ "learning_rate": 0.002973945746391701,
+ "loss": 6.166373252868652,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.012915392406284809,
+ "learning_rate": 0.0029736965673321032,
+ "loss": 6.186316013336182,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.010772830806672573,
+ "learning_rate": 0.0029734462130270273,
+ "loss": 6.172578811645508,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.010274911299347878,
+ "learning_rate": 0.002973194683678181,
+ "loss": 6.206851005554199,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.013210494071245193,
+ "learning_rate": 0.0029729419794882164,
+ "loss": 6.21115779876709,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.013372313231229782,
+ "learning_rate": 0.0029726881006607326,
+ "loss": 6.18442440032959,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.008105908520519733,
+ "learning_rate": 0.0029724330474002743,
+ "loss": 6.204354286193848,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.014167108573019505,
+ "learning_rate": 0.0029721768199123346,
+ "loss": 6.180662155151367,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.008339803665876389,
+ "learning_rate": 0.002971919418403351,
+ "loss": 6.150079727172852,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.015939386561512947,
+ "learning_rate": 0.0029716608430807073,
+ "loss": 6.175203323364258,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.010777109302580357,
+ "learning_rate": 0.0029714010941527327,
+ "loss": 6.166106224060059,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.011274735443294048,
+ "learning_rate": 0.002971140171828702,
+ "loss": 6.163111686706543,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013626808300614357,
+ "learning_rate": 0.002970878076318836,
+ "loss": 6.140850067138672,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.016907762736082077,
+ "learning_rate": 0.0029706148078343,
+ "loss": 6.192012786865234,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.018817441537976265,
+ "learning_rate": 0.002970350366587205,
+ "loss": 6.162238121032715,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.02829768694937229,
+ "learning_rate": 0.0029700847527906068,
+ "loss": 6.159066200256348,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.02034052088856697,
+ "learning_rate": 0.002969817966658505,
+ "loss": 6.164158344268799,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.01816369965672493,
+ "learning_rate": 0.0029695500084058447,
+ "loss": 6.132316589355469,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.013685261830687523,
+ "learning_rate": 0.0029692808782485146,
+ "loss": 6.152181625366211,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.009791149757802486,
+ "learning_rate": 0.0029690105764033485,
+ "loss": 6.138544082641602,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.014827069826424122,
+ "learning_rate": 0.002968739103088124,
+ "loss": 6.135766983032227,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.01391249243170023,
+ "learning_rate": 0.0029684664585215618,
+ "loss": 6.146172523498535,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.012729899026453495,
+ "learning_rate": 0.002968192642923326,
+ "loss": 6.122613430023193,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.014175458811223507,
+ "learning_rate": 0.002967917656514026,
+ "loss": 6.12574577331543,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.010958840139210224,
+ "learning_rate": 0.002967641499515213,
+ "loss": 6.111088752746582,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.015380816534161568,
+ "learning_rate": 0.0029673641721493817,
+ "loss": 6.100615501403809,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.025339096784591675,
+ "learning_rate": 0.00296708567463997,
+ "loss": 6.139659881591797,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.03695995733141899,
+ "learning_rate": 0.002966806007211358,
+ "loss": 6.160996437072754,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.0408046580851078,
+ "learning_rate": 0.0029665251700888686,
+ "loss": 6.173980712890625,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.025022024288773537,
+ "learning_rate": 0.0029662431634987674,
+ "loss": 6.145794868469238,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.027254154905676842,
+ "learning_rate": 0.002965959987668262,
+ "loss": 6.16928243637085,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.06191903352737427,
+ "learning_rate": 0.0029656756428255017,
+ "loss": 6.138525009155273,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0348956398665905,
+ "learning_rate": 0.002965390129199579,
+ "loss": 6.185477256774902,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.020859677344560623,
+ "learning_rate": 0.002965103447020526,
+ "loss": 6.148846626281738,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.022810814902186394,
+ "learning_rate": 0.0029648155965193174,
+ "loss": 6.1701178550720215,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.017981817945837975,
+ "learning_rate": 0.00296452657792787,
+ "loss": 6.1323771476745605,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.014905866235494614,
+ "learning_rate": 0.00296423639147904,
+ "loss": 6.123426914215088,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.01463468000292778,
+ "learning_rate": 0.002963945037406626,
+ "loss": 6.146122932434082,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.021616583690047264,
+ "learning_rate": 0.002963652515945366,
+ "loss": 6.093459606170654,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.01550968922674656,
+ "learning_rate": 0.00296335882733094,
+ "loss": 6.114929676055908,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.014597428031265736,
+ "learning_rate": 0.0029630639717999675,
+ "loss": 6.113945960998535,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.018249280750751495,
+ "learning_rate": 0.002962767949590008,
+ "loss": 6.130507469177246,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.014160148799419403,
+ "learning_rate": 0.0029624707609395614,
+ "loss": 6.086946487426758,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.015563832595944405,
+ "learning_rate": 0.002962172406088067,
+ "loss": 6.083332538604736,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.012240015901625156,
+ "learning_rate": 0.002961872885275904,
+ "loss": 6.0822906494140625,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.0141651201993227,
+ "learning_rate": 0.002961572198744391,
+ "loss": 6.049562454223633,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.010004645213484764,
+ "learning_rate": 0.0029612703467357866,
+ "loss": 6.067063808441162,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.011790835298597813,
+ "learning_rate": 0.0029609673294932856,
+ "loss": 6.062193870544434,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.011088060215115547,
+ "learning_rate": 0.002960663147261025,
+ "loss": 6.070950031280518,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.010387620888650417,
+ "learning_rate": 0.002960357800284078,
+ "loss": 6.032423973083496,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.012074458412826061,
+ "learning_rate": 0.002960051288808458,
+ "loss": 6.066788673400879,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.097111701965332,
+ "eval_runtime": 51.8742,
+ "eval_samples_per_second": 47.075,
+ "eval_steps_per_second": 1.484,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.013429952785372734,
+ "learning_rate": 0.0029597436130811153,
+ "loss": 6.022172451019287,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.01764705777168274,
+ "learning_rate": 0.0029594347733499387,
+ "loss": 6.029745101928711,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.02403099089860916,
+ "learning_rate": 0.002959124769863755,
+ "loss": 6.04707145690918,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.029030559584498405,
+ "learning_rate": 0.002958813602872328,
+ "loss": 6.05479097366333,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.029980644583702087,
+ "learning_rate": 0.0029585012726263595,
+ "loss": 6.068657875061035,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.021424276754260063,
+ "learning_rate": 0.0029581877793774885,
+ "loss": 6.062008857727051,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.018076395615935326,
+ "learning_rate": 0.002957873123378291,
+ "loss": 6.023279190063477,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.017688792198896408,
+ "learning_rate": 0.002957557304882279,
+ "loss": 6.014641284942627,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.02313626930117607,
+ "learning_rate": 0.0029572403241439026,
+ "loss": 6.026420593261719,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.04045062139630318,
+ "learning_rate": 0.0029569221814185467,
+ "loss": 6.007199764251709,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.015792209655046463,
+ "learning_rate": 0.0029566028769625334,
+ "loss": 6.017209053039551,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.01990479789674282,
+ "learning_rate": 0.002956282411033121,
+ "loss": 5.982841491699219,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.01191043108701706,
+ "learning_rate": 0.002955960783888503,
+ "loss": 6.0035929679870605,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.01732831820845604,
+ "learning_rate": 0.002955637995787807,
+ "loss": 6.025006294250488,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.032948266714811325,
+ "learning_rate": 0.0029553140469910995,
+ "loss": 6.007736682891846,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.011192727833986282,
+ "learning_rate": 0.00295498893775938,
+ "loss": 6.0022687911987305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014646840281784534,
+ "learning_rate": 0.0029546626683545824,
+ "loss": 5.976479530334473,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01589796133339405,
+ "learning_rate": 0.002954335239039576,
+ "loss": 5.980949878692627,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.015129734762012959,
+ "learning_rate": 0.0029540066500781657,
+ "loss": 6.0073394775390625,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.011140518821775913,
+ "learning_rate": 0.002953676901735088,
+ "loss": 5.976587295532227,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.0109480582177639,
+ "learning_rate": 0.0029533459942760174,
+ "loss": 6.001427173614502,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.013346790336072445,
+ "learning_rate": 0.0029530139279675583,
+ "loss": 5.96919059753418,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.012474581599235535,
+ "learning_rate": 0.002952680703077251,
+ "loss": 5.970426559448242,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01140566635876894,
+ "learning_rate": 0.002952346319873569,
+ "loss": 5.989171504974365,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.0137029392644763,
+ "learning_rate": 0.0029520107786259187,
+ "loss": 6.010115623474121,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015517761930823326,
+ "learning_rate": 0.002951674079604639,
+ "loss": 6.003649711608887,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.02173629403114319,
+ "learning_rate": 0.0029513362230810035,
+ "loss": 5.967074871063232,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02683914266526699,
+ "learning_rate": 0.002950997209327216,
+ "loss": 5.994434833526611,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.025865858420729637,
+ "learning_rate": 0.002950657038616415,
+ "loss": 6.010523796081543,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02544897235929966,
+ "learning_rate": 0.002950315711222668,
+ "loss": 5.994078159332275,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.028426019474864006,
+ "learning_rate": 0.0029499732274209777,
+ "loss": 6.020925521850586,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.026174750179052353,
+ "learning_rate": 0.0029496295874872768,
+ "loss": 5.985616683959961,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.02802138403058052,
+ "learning_rate": 0.00294928479169843,
+ "loss": 5.973659038543701,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.019348938018083572,
+ "learning_rate": 0.002948938840332233,
+ "loss": 5.990137577056885,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.016294041648507118,
+ "learning_rate": 0.0029485917336674115,
+ "loss": 5.962082862854004,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.016167785972356796,
+ "learning_rate": 0.002948243471983625,
+ "loss": 5.9594573974609375,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.014119232073426247,
+ "learning_rate": 0.00294789405556146,
+ "loss": 5.894467353820801,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.013805347494781017,
+ "learning_rate": 0.002947543484682436,
+ "loss": 5.964014530181885,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.014496497809886932,
+ "learning_rate": 0.002947191759629001,
+ "loss": 5.989920616149902,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.014813800342381,
+ "learning_rate": 0.0029468388806845333,
+ "loss": 5.9669413566589355,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.01153622381389141,
+ "learning_rate": 0.0029464848481333423,
+ "loss": 5.918018341064453,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.010525977239012718,
+ "learning_rate": 0.0029461296622606647,
+ "loss": 5.892477989196777,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.011410463601350784,
+ "learning_rate": 0.002945773323352667,
+ "loss": 5.925249099731445,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.010949424467980862,
+ "learning_rate": 0.0029454158316964464,
+ "loss": 5.90645694732666,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.00962804164737463,
+ "learning_rate": 0.002945057187580026,
+ "loss": 5.917054176330566,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.011517546139657497,
+ "learning_rate": 0.0029446973912923597,
+ "loss": 5.919445514678955,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01567666605114937,
+ "learning_rate": 0.0029443364431233285,
+ "loss": 5.884111404418945,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.016781140118837357,
+ "learning_rate": 0.002943974343363742,
+ "loss": 5.9252119064331055,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.02244546078145504,
+ "learning_rate": 0.0029436110923053375,
+ "loss": 5.933143615722656,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.028186792507767677,
+ "learning_rate": 0.00294324669024078,
+ "loss": 5.9079132080078125,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.028585223481059074,
+ "learning_rate": 0.002942881137463661,
+ "loss": 5.925262451171875,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.02781485952436924,
+ "learning_rate": 0.0029425144342685003,
+ "loss": 5.963315010070801,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.021486928686499596,
+ "learning_rate": 0.002942146580950744,
+ "loss": 5.930115699768066,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.016890818253159523,
+ "learning_rate": 0.002941777577806765,
+ "loss": 5.921439170837402,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.0169126745313406,
+ "learning_rate": 0.0029414074251338618,
+ "loss": 5.928220272064209,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.014659176580607891,
+ "learning_rate": 0.002941036123230261,
+ "loss": 5.934546947479248,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.015337883494794369,
+ "learning_rate": 0.002940663672395112,
+ "loss": 5.890145301818848,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01254988368600607,
+ "learning_rate": 0.002940290072928493,
+ "loss": 5.875492095947266,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.014658362604677677,
+ "learning_rate": 0.0029399153251314066,
+ "loss": 5.901272773742676,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.017342250794172287,
+ "learning_rate": 0.002939539429305779,
+ "loss": 5.875853538513184,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.016825608909130096,
+ "learning_rate": 0.0029391623857544643,
+ "loss": 5.941972732543945,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.016716884449124336,
+ "learning_rate": 0.002938784194781238,
+ "loss": 5.912367343902588,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.021707067266106606,
+ "learning_rate": 0.002938404856690803,
+ "loss": 5.895319938659668,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.030519135296344757,
+ "learning_rate": 0.002938024371788785,
+ "loss": 5.897165298461914,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.02572188340127468,
+ "learning_rate": 0.002937642740381733,
+ "loss": 5.879164695739746,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.016286717727780342,
+ "learning_rate": 0.002937259962777122,
+ "loss": 5.841421127319336,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.020656082779169083,
+ "learning_rate": 0.002936876039283347,
+ "loss": 5.9222259521484375,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.015350459143519402,
+ "learning_rate": 0.0029364909702097298,
+ "loss": 5.885847091674805,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01708570495247841,
+ "learning_rate": 0.0029361047558665136,
+ "loss": 5.872044563293457,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.013537376187741756,
+ "learning_rate": 0.0029357173965648634,
+ "loss": 5.866434097290039,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.013208637945353985,
+ "learning_rate": 0.002935328892616868,
+ "loss": 5.8624491691589355,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.012249554507434368,
+ "learning_rate": 0.002934939244335538,
+ "loss": 5.833232879638672,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.90985107421875,
+ "eval_runtime": 51.0696,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.01382362935692072,
+ "learning_rate": 0.0029345484520348062,
+ "loss": 5.835695266723633,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.012947693467140198,
+ "learning_rate": 0.0029341565160295262,
+ "loss": 5.826190948486328,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.013653872534632683,
+ "learning_rate": 0.002933763436635475,
+ "loss": 5.831789016723633,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.027596980333328247,
+ "learning_rate": 0.0029333692141693484,
+ "loss": 5.839963436126709,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01637990213930607,
+ "learning_rate": 0.0029329738489487648,
+ "loss": 5.805248260498047,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.02169182151556015,
+ "learning_rate": 0.0029325773412922627,
+ "loss": 5.838248252868652,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.03518201410770416,
+ "learning_rate": 0.0029321796915193015,
+ "loss": 5.834336280822754,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04983877018094063,
+ "learning_rate": 0.0029317808999502603,
+ "loss": 5.879856109619141,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05085664615035057,
+ "learning_rate": 0.0029313809669064377,
+ "loss": 5.946531295776367,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.03314858302474022,
+ "learning_rate": 0.0029309798927100532,
+ "loss": 5.909746170043945,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.034118980169296265,
+ "learning_rate": 0.0029305776776842453,
+ "loss": 5.903947830200195,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.03401247411966324,
+ "learning_rate": 0.0029301743221530705,
+ "loss": 5.936162948608398,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.043827857822179794,
+ "learning_rate": 0.0029297698264415047,
+ "loss": 5.953949451446533,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.06281330436468124,
+ "learning_rate": 0.002929364190875444,
+ "loss": 5.988225936889648,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.07216024398803711,
+ "learning_rate": 0.0029289574157817014,
+ "loss": 6.187741279602051,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.055205658078193665,
+ "learning_rate": 0.002928549501488008,
+ "loss": 6.177831649780273,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.04834286496043205,
+ "learning_rate": 0.0029281404483230133,
+ "loss": 6.209604263305664,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.0447777584195137,
+ "learning_rate": 0.0029277302566162837,
+ "loss": 6.146507263183594,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.05729430541396141,
+ "learning_rate": 0.0029273189266983035,
+ "loss": 6.280686378479004,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.0410807766020298,
+ "learning_rate": 0.002926906458900473,
+ "loss": 6.246824264526367,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.044607777148485184,
+ "learning_rate": 0.0029264928535551104,
+ "loss": 6.19956111907959,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.05141043663024902,
+ "learning_rate": 0.0029260781109954515,
+ "loss": 6.118578910827637,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.029248764738440514,
+ "learning_rate": 0.0029256622315556448,
+ "loss": 6.155360221862793,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.028403814882040024,
+ "learning_rate": 0.0029252452155707582,
+ "loss": 6.097352027893066,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.025872666388750076,
+ "learning_rate": 0.002924827063376773,
+ "loss": 6.072485446929932,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.018447555601596832,
+ "learning_rate": 0.0029244077753105874,
+ "loss": 6.077045440673828,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.023166058585047722,
+ "learning_rate": 0.0029239873517100147,
+ "loss": 6.040050983428955,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.019579261541366577,
+ "learning_rate": 0.002923565792913782,
+ "loss": 6.039424896240234,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.018369073048233986,
+ "learning_rate": 0.002923143099261531,
+ "loss": 6.021180152893066,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.014188461937010288,
+ "learning_rate": 0.0029227192710938195,
+ "loss": 5.9668049812316895,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.019567398354411125,
+ "learning_rate": 0.0029222943087521182,
+ "loss": 5.989189147949219,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.014190621674060822,
+ "learning_rate": 0.0029218682125788102,
+ "loss": 5.942726135253906,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01375526562333107,
+ "learning_rate": 0.002921440982917195,
+ "loss": 5.942777633666992,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.014203300699591637,
+ "learning_rate": 0.0029210126201114836,
+ "loss": 5.927928924560547,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.014231069944798946,
+ "learning_rate": 0.0029205831245067994,
+ "loss": 5.923510551452637,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.012771515175700188,
+ "learning_rate": 0.0029201524964491796,
+ "loss": 5.908803939819336,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.013399476185441017,
+ "learning_rate": 0.0029197207362855735,
+ "loss": 5.890079021453857,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.025862423703074455,
+ "learning_rate": 0.0029192878443638427,
+ "loss": 5.919907569885254,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.011315986514091492,
+ "learning_rate": 0.0029188538210327595,
+ "loss": 5.901529788970947,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.01235947571694851,
+ "learning_rate": 0.00291841866664201,
+ "loss": 5.903595924377441,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.012175127863883972,
+ "learning_rate": 0.0029179823815421893,
+ "loss": 5.868306636810303,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01195445004850626,
+ "learning_rate": 0.0029175449660848052,
+ "loss": 5.899898529052734,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.012801481410861015,
+ "learning_rate": 0.0029171064206222745,
+ "loss": 5.890621662139893,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.011909718625247478,
+ "learning_rate": 0.002916666745507926,
+ "loss": 5.86421012878418,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01162486057728529,
+ "learning_rate": 0.0029162259410959977,
+ "loss": 5.869139671325684,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.010923407040536404,
+ "learning_rate": 0.0029157840077416374,
+ "loss": 5.869789123535156,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.011943561024963856,
+ "learning_rate": 0.002915340945800904,
+ "loss": 5.873437881469727,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.010295520536601543,
+ "learning_rate": 0.002914896755630763,
+ "loss": 5.819953918457031,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.0093373479321599,
+ "learning_rate": 0.002914451437589092,
+ "loss": 5.829481601715088,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.01182442344725132,
+ "learning_rate": 0.002914004992034674,
+ "loss": 5.855249404907227,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.00939213763922453,
+ "learning_rate": 0.0029135574193272034,
+ "loss": 5.83876371383667,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.011648556217551231,
+ "learning_rate": 0.0029131087198272813,
+ "loss": 5.824906826019287,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.011227169074118137,
+ "learning_rate": 0.002912658893896416,
+ "loss": 5.847869873046875,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.011084851808845997,
+ "learning_rate": 0.002912207941897025,
+ "loss": 5.839183807373047,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.009923825040459633,
+ "learning_rate": 0.002911755864192432,
+ "loss": 5.818821907043457,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.011300318874418736,
+ "learning_rate": 0.002911302661146867,
+ "loss": 5.819517135620117,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.009915645234286785,
+ "learning_rate": 0.0029108483331254684,
+ "loss": 5.776939392089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.008965205401182175,
+ "learning_rate": 0.00291039288049428,
+ "loss": 5.824148178100586,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.011025927029550076,
+ "learning_rate": 0.0029099363036202515,
+ "loss": 5.832704544067383,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.010143411345779896,
+ "learning_rate": 0.0029094786028712387,
+ "loss": 5.802353382110596,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.010021256282925606,
+ "learning_rate": 0.002909019778616003,
+ "loss": 5.802136421203613,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.011356706731021404,
+ "learning_rate": 0.002908559831224211,
+ "loss": 5.807962417602539,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.010525298304855824,
+ "learning_rate": 0.002908098761066433,
+ "loss": 5.787661552429199,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.011147079057991505,
+ "learning_rate": 0.0029076365685141463,
+ "loss": 5.801015377044678,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.008777810260653496,
+ "learning_rate": 0.00290717325393973,
+ "loss": 5.779858589172363,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.011423609219491482,
+ "learning_rate": 0.002906708817716469,
+ "loss": 5.777166843414307,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.008639856241643429,
+ "learning_rate": 0.00290624326021855,
+ "loss": 5.747189521789551,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.009628898464143276,
+ "learning_rate": 0.002905776581821066,
+ "loss": 5.7666916847229,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.011988377198576927,
+ "learning_rate": 0.0029053087829000097,
+ "loss": 5.777951240539551,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.01693224348127842,
+ "learning_rate": 0.00290483986383228,
+ "loss": 5.818393707275391,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.020264796912670135,
+ "learning_rate": 0.002904369824995674,
+ "loss": 5.747611045837402,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.018283747136592865,
+ "learning_rate": 0.0029038986667688964,
+ "loss": 5.789663314819336,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.825676918029785,
+ "eval_runtime": 57.6908,
+ "eval_samples_per_second": 42.329,
+ "eval_steps_per_second": 1.335,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.011447477154433727,
+ "learning_rate": 0.002903426389531549,
+ "loss": 5.724975109100342,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.012539781630039215,
+ "learning_rate": 0.002902952993664138,
+ "loss": 5.747196674346924,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.012179837562143803,
+ "learning_rate": 0.002902478479548069,
+ "loss": 5.748995780944824,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.019896456971764565,
+ "learning_rate": 0.00290200284756565,
+ "loss": 5.751996994018555,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.022261742502450943,
+ "learning_rate": 0.002901526098100089,
+ "loss": 5.7516984939575195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02683536894619465,
+ "learning_rate": 0.0029010482315354946,
+ "loss": 5.716037750244141,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.018057072535157204,
+ "learning_rate": 0.0029005692482568747,
+ "loss": 5.778608322143555,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.011646891944110394,
+ "learning_rate": 0.002900089148650138,
+ "loss": 5.7308268547058105,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02230495773255825,
+ "learning_rate": 0.002899607933102092,
+ "loss": 5.768552780151367,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.01212464738637209,
+ "learning_rate": 0.002899125602000443,
+ "loss": 5.733015060424805,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.011764907278120518,
+ "learning_rate": 0.0028986421557337973,
+ "loss": 5.748104095458984,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.01593147963285446,
+ "learning_rate": 0.002898157594691658,
+ "loss": 5.701960563659668,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.016333773732185364,
+ "learning_rate": 0.0028976719192644277,
+ "loss": 5.716808795928955,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.017547868192195892,
+ "learning_rate": 0.002897185129843406,
+ "loss": 5.724377632141113,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.01662801019847393,
+ "learning_rate": 0.002896697226820791,
+ "loss": 5.737432956695557,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.015781691297888756,
+ "learning_rate": 0.002896208210589676,
+ "loss": 5.7376203536987305,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.014363214373588562,
+ "learning_rate": 0.0028957180815440534,
+ "loss": 5.7342610359191895,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.015048054046928883,
+ "learning_rate": 0.0028952268400788115,
+ "loss": 5.715073585510254,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.013753226026892662,
+ "learning_rate": 0.002894734486589734,
+ "loss": 5.73264217376709,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.015562563203275204,
+ "learning_rate": 0.002894241021473502,
+ "loss": 5.7072038650512695,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.01520819216966629,
+ "learning_rate": 0.0028937464451276916,
+ "loss": 5.683300018310547,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.013024724088609219,
+ "learning_rate": 0.002893250757950773,
+ "loss": 5.726273536682129,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.012760856188833714,
+ "learning_rate": 0.0028927539603421128,
+ "loss": 5.7249064445495605,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.013542579486966133,
+ "learning_rate": 0.0028922560527019723,
+ "loss": 5.70849084854126,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.014801754616200924,
+ "learning_rate": 0.002891757035431506,
+ "loss": 5.725185394287109,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.013000023551285267,
+ "learning_rate": 0.0028912569089327637,
+ "loss": 5.705799102783203,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.012919292785227299,
+ "learning_rate": 0.002890755673608688,
+ "loss": 5.674324989318848,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.015904201194643974,
+ "learning_rate": 0.002890253329863115,
+ "loss": 5.680788993835449,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.015467721968889236,
+ "learning_rate": 0.0028897498781007744,
+ "loss": 5.707531929016113,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.015157260000705719,
+ "learning_rate": 0.002889245318727288,
+ "loss": 5.689894676208496,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.014204306527972221,
+ "learning_rate": 0.00288873965214917,
+ "loss": 5.713001251220703,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.022045845165848732,
+ "learning_rate": 0.0028882328787738273,
+ "loss": 5.657631874084473,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.013380175456404686,
+ "learning_rate": 0.0028877249990095573,
+ "loss": 5.702192306518555,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.016698408871889114,
+ "learning_rate": 0.0028872160132655503,
+ "loss": 5.695206642150879,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.020001472905278206,
+ "learning_rate": 0.0028867059219518866,
+ "loss": 5.731742858886719,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.027350103482604027,
+ "learning_rate": 0.0028861947254795375,
+ "loss": 5.712314605712891,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.021862946450710297,
+ "learning_rate": 0.002885682424260365,
+ "loss": 5.697043418884277,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.019216977059841156,
+ "learning_rate": 0.0028851690187071204,
+ "loss": 5.698813438415527,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.02131095714867115,
+ "learning_rate": 0.002884654509233446,
+ "loss": 5.6918864250183105,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.016583887860178947,
+ "learning_rate": 0.0028841388962538724,
+ "loss": 5.711456298828125,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.016998684033751488,
+ "learning_rate": 0.0028836221801838194,
+ "loss": 5.675747871398926,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.01705755852162838,
+ "learning_rate": 0.0028831043614395966,
+ "loss": 5.697177410125732,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.014604352414608002,
+ "learning_rate": 0.0028825854404384,
+ "loss": 5.681950092315674,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.013561096973717213,
+ "learning_rate": 0.0028820654175983162,
+ "loss": 5.67357063293457,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.021149173378944397,
+ "learning_rate": 0.0028815442933383174,
+ "loss": 5.709477424621582,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.01431041955947876,
+ "learning_rate": 0.0028810220680782644,
+ "loss": 5.7012200355529785,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.01429709978401661,
+ "learning_rate": 0.0028804987422389045,
+ "loss": 5.689794540405273,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.032025713473558426,
+ "learning_rate": 0.0028799743162418712,
+ "loss": 5.6713762283325195,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.013168714940547943,
+ "learning_rate": 0.002879448790509686,
+ "loss": 5.685478210449219,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.014088666997849941,
+ "learning_rate": 0.0028789221654657556,
+ "loss": 5.645514965057373,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.013320053927600384,
+ "learning_rate": 0.002878394441534372,
+ "loss": 5.678678512573242,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.022180087864398956,
+ "learning_rate": 0.002877865619140712,
+ "loss": 5.660525321960449,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.01271889265626669,
+ "learning_rate": 0.002877335698710839,
+ "loss": 5.652164459228516,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.011617431417107582,
+ "learning_rate": 0.002876804680671701,
+ "loss": 5.6389946937561035,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.012040726840496063,
+ "learning_rate": 0.0028762725654511282,
+ "loss": 5.628744602203369,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.013148069381713867,
+ "learning_rate": 0.0028757393534778377,
+ "loss": 5.69610595703125,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.01826276071369648,
+ "learning_rate": 0.0028752050451814276,
+ "loss": 5.634238243103027,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.020393460988998413,
+ "learning_rate": 0.0028746696409923805,
+ "loss": 5.656968593597412,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.017072148621082306,
+ "learning_rate": 0.0028741331413420627,
+ "loss": 5.668219089508057,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.01694098301231861,
+ "learning_rate": 0.0028735955466627217,
+ "loss": 5.670073509216309,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.022760670632123947,
+ "learning_rate": 0.002873056857387488,
+ "loss": 5.649288177490234,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.03613442927598953,
+ "learning_rate": 0.0028725170739503734,
+ "loss": 5.723211288452148,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.04232329502701759,
+ "learning_rate": 0.002871976196786272,
+ "loss": 5.757920742034912,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.044210851192474365,
+ "learning_rate": 0.002871434226330959,
+ "loss": 5.702426910400391,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.044886402785778046,
+ "learning_rate": 0.0028708911630210905,
+ "loss": 5.742881774902344,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.022805195301771164,
+ "learning_rate": 0.002870347007294202,
+ "loss": 5.716353416442871,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02219395525753498,
+ "learning_rate": 0.00286980175958871,
+ "loss": 5.68503475189209,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.019240057095885277,
+ "learning_rate": 0.002869255420343911,
+ "loss": 5.700656890869141,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.01957295648753643,
+ "learning_rate": 0.0028687079899999807,
+ "loss": 5.70595645904541,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016326555982232094,
+ "learning_rate": 0.0028681594689979736,
+ "loss": 5.697229385375977,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015815885737538338,
+ "learning_rate": 0.0028676098577798234,
+ "loss": 5.7020792961120605,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.015865175053477287,
+ "learning_rate": 0.002867059156788342,
+ "loss": 5.648272514343262,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.7417449951171875,
+ "eval_runtime": 51.2263,
+ "eval_samples_per_second": 47.671,
+ "eval_steps_per_second": 1.503,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.016063183546066284,
+ "learning_rate": 0.0028665073664672185,
+ "loss": 5.65411376953125,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.014323881827294827,
+ "learning_rate": 0.0028659544872610214,
+ "loss": 5.629649639129639,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.014005238190293312,
+ "learning_rate": 0.002865400519615195,
+ "loss": 5.655135631561279,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.01385372132062912,
+ "learning_rate": 0.0028648454639760615,
+ "loss": 5.632213115692139,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.014073756523430347,
+ "learning_rate": 0.0028642893207908187,
+ "loss": 5.626431465148926,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.019551943987607956,
+ "learning_rate": 0.0028637320905075406,
+ "loss": 5.595447063446045,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.017571093514561653,
+ "learning_rate": 0.002863173773575179,
+ "loss": 5.661581039428711,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.01691303588449955,
+ "learning_rate": 0.0028626143704435585,
+ "loss": 5.641129493713379,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.01584242656826973,
+ "learning_rate": 0.0028620538815633807,
+ "loss": 5.640626907348633,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.014592235907912254,
+ "learning_rate": 0.0028614923073862213,
+ "loss": 5.611204147338867,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.016636405140161514,
+ "learning_rate": 0.0028609296483645304,
+ "loss": 5.614716053009033,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.01370833721011877,
+ "learning_rate": 0.002860365904951632,
+ "loss": 5.633011817932129,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.013218311592936516,
+ "learning_rate": 0.002859801077601725,
+ "loss": 5.588958740234375,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.016971616074442863,
+ "learning_rate": 0.0028592351667698786,
+ "loss": 5.635064125061035,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.015524905174970627,
+ "learning_rate": 0.0028586681729120393,
+ "loss": 5.611454963684082,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.014769157394766808,
+ "learning_rate": 0.0028581000964850215,
+ "loss": 5.62175178527832,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.014313343912363052,
+ "learning_rate": 0.002857530937946516,
+ "loss": 5.603628158569336,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.016748322173953056,
+ "learning_rate": 0.002856960697755083,
+ "loss": 5.617765426635742,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.013503300957381725,
+ "learning_rate": 0.0028563893763701538,
+ "loss": 5.636135578155518,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.012275655753910542,
+ "learning_rate": 0.002855816974252033,
+ "loss": 5.552533149719238,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.01288773212581873,
+ "learning_rate": 0.002855243491861894,
+ "loss": 5.620527267456055,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.013687008060514927,
+ "learning_rate": 0.002854668929661781,
+ "loss": 5.618595123291016,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.015678439289331436,
+ "learning_rate": 0.0028540932881146092,
+ "loss": 5.604031562805176,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.013567988760769367,
+ "learning_rate": 0.002853516567684162,
+ "loss": 5.549149513244629,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0157297495752573,
+ "learning_rate": 0.002852938768835092,
+ "loss": 5.618505477905273,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.011867019347846508,
+ "learning_rate": 0.0028523598920329233,
+ "loss": 5.579049110412598,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.01398563478142023,
+ "learning_rate": 0.0028517799377440443,
+ "loss": 5.597678184509277,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.015525273978710175,
+ "learning_rate": 0.0028511989064357148,
+ "loss": 5.601663589477539,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.01588466763496399,
+ "learning_rate": 0.002850616798576061,
+ "loss": 5.5650434494018555,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.014276969246566296,
+ "learning_rate": 0.0028500336146340772,
+ "loss": 5.607125282287598,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.013253689743578434,
+ "learning_rate": 0.002849449355079623,
+ "loss": 5.6299824714660645,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.01176661066710949,
+ "learning_rate": 0.0028488640203834263,
+ "loss": 5.560839653015137,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.011712565086781979,
+ "learning_rate": 0.0028482776110170813,
+ "loss": 5.618312835693359,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.013515647500753403,
+ "learning_rate": 0.002847690127453047,
+ "loss": 5.566764831542969,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.020177524536848068,
+ "learning_rate": 0.0028471015701646485,
+ "loss": 5.554571628570557,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.029197178781032562,
+ "learning_rate": 0.0028465119396260757,
+ "loss": 5.6097259521484375,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.027947140857577324,
+ "learning_rate": 0.0028459212363123824,
+ "loss": 5.65598201751709,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.021089809015393257,
+ "learning_rate": 0.0028453294606994887,
+ "loss": 5.67195987701416,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.019936271011829376,
+ "learning_rate": 0.0028447366132641773,
+ "loss": 5.609482765197754,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.0176444910466671,
+ "learning_rate": 0.002844142694484094,
+ "loss": 5.60432767868042,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.023154428228735924,
+ "learning_rate": 0.0028435477048377495,
+ "loss": 5.616270065307617,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.018391480669379234,
+ "learning_rate": 0.002842951644804515,
+ "loss": 5.600588321685791,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.013716557994484901,
+ "learning_rate": 0.002842354514864627,
+ "loss": 5.595095634460449,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.014372860081493855,
+ "learning_rate": 0.002841756315499181,
+ "loss": 5.584517478942871,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.014930845238268375,
+ "learning_rate": 0.002841157047190135,
+ "loss": 5.587005138397217,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.012631502002477646,
+ "learning_rate": 0.0028405567104203095,
+ "loss": 5.594094276428223,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.014677002094686031,
+ "learning_rate": 0.0028399553056733854,
+ "loss": 5.569582939147949,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.015558961778879166,
+ "learning_rate": 0.0028393528334339025,
+ "loss": 5.596535682678223,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.014322095550596714,
+ "learning_rate": 0.0028387492941872624,
+ "loss": 5.570324420928955,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.012913414277136326,
+ "learning_rate": 0.002838144688419726,
+ "loss": 5.573002338409424,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.01251310296356678,
+ "learning_rate": 0.0028375390166184137,
+ "loss": 5.550327777862549,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.013847017660737038,
+ "learning_rate": 0.0028369322792713026,
+ "loss": 5.54567813873291,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.013789358548820019,
+ "learning_rate": 0.0028363244768672313,
+ "loss": 5.571495056152344,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.01236631814390421,
+ "learning_rate": 0.0028357156098958954,
+ "loss": 5.569730758666992,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.013210169039666653,
+ "learning_rate": 0.002835105678847847,
+ "loss": 5.53984260559082,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.01604137383401394,
+ "learning_rate": 0.0028344946842144975,
+ "loss": 5.530598163604736,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.01708815060555935,
+ "learning_rate": 0.002833882626488114,
+ "loss": 5.529256820678711,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.017222166061401367,
+ "learning_rate": 0.0028332695061618197,
+ "loss": 5.564949035644531,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.01803218014538288,
+ "learning_rate": 0.002832655323729595,
+ "loss": 5.560049057006836,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.01658225618302822,
+ "learning_rate": 0.002832040079686276,
+ "loss": 5.553183555603027,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.016834143549203873,
+ "learning_rate": 0.002831423774527554,
+ "loss": 5.573678970336914,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.01856408454477787,
+ "learning_rate": 0.002830806408749974,
+ "loss": 5.539497375488281,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.026909375563263893,
+ "learning_rate": 0.002830187982850936,
+ "loss": 5.551433563232422,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.016046348959207535,
+ "learning_rate": 0.002829568497328696,
+ "loss": 5.566728591918945,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.018582617864012718,
+ "learning_rate": 0.0028289479526823614,
+ "loss": 5.543226718902588,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.0211537666618824,
+ "learning_rate": 0.0028283263494118943,
+ "loss": 5.560711860656738,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.013894101604819298,
+ "learning_rate": 0.0028277036880181095,
+ "loss": 5.528890132904053,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.017581885680556297,
+ "learning_rate": 0.0028270799690026725,
+ "loss": 5.57473087310791,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.01622696965932846,
+ "learning_rate": 0.002826455192868104,
+ "loss": 5.578961372375488,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.01966337114572525,
+ "learning_rate": 0.0028258293601177743,
+ "loss": 5.513942718505859,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021581366658210754,
+ "learning_rate": 0.002825202471255906,
+ "loss": 5.573669910430908,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.020468758419156075,
+ "learning_rate": 0.002824574526787571,
+ "loss": 5.516103744506836,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.626922130584717,
+ "eval_runtime": 54.2404,
+ "eval_samples_per_second": 45.022,
+ "eval_steps_per_second": 1.42,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.014597049914300442,
+ "learning_rate": 0.0028239455272186936,
+ "loss": 5.506453514099121,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.020277054980397224,
+ "learning_rate": 0.002823315473056048,
+ "loss": 5.535350799560547,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.019103925675153732,
+ "learning_rate": 0.0028226843648072562,
+ "loss": 5.522078037261963,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.07158710062503815,
+ "learning_rate": 0.0028220522029807913,
+ "loss": 5.531142234802246,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01958007924258709,
+ "learning_rate": 0.0028214189880859746,
+ "loss": 5.484055042266846,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.0185184795409441,
+ "learning_rate": 0.002820784720632976,
+ "loss": 5.511264801025391,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01976991444826126,
+ "learning_rate": 0.0028201494011328133,
+ "loss": 5.51821231842041,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.02188604325056076,
+ "learning_rate": 0.002819513030097352,
+ "loss": 5.483095169067383,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.019405223429203033,
+ "learning_rate": 0.002818875608039305,
+ "loss": 5.539701461791992,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.016907615587115288,
+ "learning_rate": 0.0028182371354722306,
+ "loss": 5.518251419067383,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.02160538360476494,
+ "learning_rate": 0.0028175976129105364,
+ "loss": 5.509786605834961,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.023125721141695976,
+ "learning_rate": 0.002816957040869473,
+ "loss": 5.535561561584473,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.022233380004763603,
+ "learning_rate": 0.0028163154198651372,
+ "loss": 5.553831100463867,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.017227552831172943,
+ "learning_rate": 0.002815672750414473,
+ "loss": 5.547080993652344,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.017122473567724228,
+ "learning_rate": 0.0028150290330352665,
+ "loss": 5.493778228759766,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.020641913637518883,
+ "learning_rate": 0.0028143842682461498,
+ "loss": 5.536645889282227,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.017383966594934464,
+ "learning_rate": 0.0028137384565665967,
+ "loss": 5.506233215332031,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.017969606444239616,
+ "learning_rate": 0.002813091598516928,
+ "loss": 5.524741172790527,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.01936613954603672,
+ "learning_rate": 0.0028124436946183047,
+ "loss": 5.460989952087402,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.017492495477199554,
+ "learning_rate": 0.0028117947453927313,
+ "loss": 5.487619876861572,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.018786396831274033,
+ "learning_rate": 0.0028111447513630545,
+ "loss": 5.516829013824463,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.020149117335677147,
+ "learning_rate": 0.0028104937130529624,
+ "loss": 5.460541725158691,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.024668103083968163,
+ "learning_rate": 0.0028098416309869857,
+ "loss": 5.497687339782715,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.015888608992099762,
+ "learning_rate": 0.002809188505690494,
+ "loss": 5.494459629058838,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.019327623769640923,
+ "learning_rate": 0.0028085343376896997,
+ "loss": 5.486351490020752,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.026855193078517914,
+ "learning_rate": 0.0028078791275116532,
+ "loss": 5.509825706481934,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.026340370997786522,
+ "learning_rate": 0.002807222875684246,
+ "loss": 5.527161598205566,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.02200835384428501,
+ "learning_rate": 0.002806565582736208,
+ "loss": 5.507146835327148,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.020420534536242485,
+ "learning_rate": 0.0028059072491971085,
+ "loss": 5.5162506103515625,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02493203990161419,
+ "learning_rate": 0.0028052478755973545,
+ "loss": 5.5252556800842285,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.021293973550200462,
+ "learning_rate": 0.0028045874624681925,
+ "loss": 5.523589134216309,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.016593236476182938,
+ "learning_rate": 0.0028039260103417045,
+ "loss": 5.500586032867432,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.018857723101973534,
+ "learning_rate": 0.002803263519750811,
+ "loss": 5.52933406829834,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.015930436551570892,
+ "learning_rate": 0.0028025999912292677,
+ "loss": 5.48813533782959,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.019427936524152756,
+ "learning_rate": 0.002801935425311669,
+ "loss": 5.46573543548584,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.018123053014278412,
+ "learning_rate": 0.0028012698225334427,
+ "loss": 5.499445915222168,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.02035410702228546,
+ "learning_rate": 0.002800603183430853,
+ "loss": 5.513540267944336,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.025838350877165794,
+ "learning_rate": 0.0027999355085409996,
+ "loss": 5.565637588500977,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.029621612280607224,
+ "learning_rate": 0.002799266798401816,
+ "loss": 5.5619215965271,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.028948379680514336,
+ "learning_rate": 0.0027985970535520685,
+ "loss": 5.551331520080566,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.022135930135846138,
+ "learning_rate": 0.0027979262745313604,
+ "loss": 5.527722358703613,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.028884069994091988,
+ "learning_rate": 0.0027972544618801256,
+ "loss": 5.554599761962891,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.02499174140393734,
+ "learning_rate": 0.00279658161613963,
+ "loss": 5.543351173400879,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.021131042391061783,
+ "learning_rate": 0.0027959077378519755,
+ "loss": 5.496039390563965,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.02957489714026451,
+ "learning_rate": 0.002795232827560092,
+ "loss": 5.552300453186035,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.022832119837403297,
+ "learning_rate": 0.0027945568858077433,
+ "loss": 5.488651275634766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.02074701525270939,
+ "learning_rate": 0.002793879913139523,
+ "loss": 5.491244792938232,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.019919803366065025,
+ "learning_rate": 0.002793201910100856,
+ "loss": 5.483855247497559,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.01858169585466385,
+ "learning_rate": 0.002792522877237998,
+ "loss": 5.50787353515625,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02085857465863228,
+ "learning_rate": 0.002791842815098032,
+ "loss": 5.512872695922852,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.01407450158149004,
+ "learning_rate": 0.002791161724228872,
+ "loss": 5.472590446472168,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.01672944612801075,
+ "learning_rate": 0.002790479605179262,
+ "loss": 5.469183921813965,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.016565700992941856,
+ "learning_rate": 0.0027897964584987713,
+ "loss": 5.493677139282227,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.017017360776662827,
+ "learning_rate": 0.0027891122847377996,
+ "loss": 5.464931488037109,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.019842486828565598,
+ "learning_rate": 0.002788427084447574,
+ "loss": 5.461293697357178,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.016210904344916344,
+ "learning_rate": 0.0027877408581801467,
+ "loss": 5.454716682434082,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.02064039371907711,
+ "learning_rate": 0.002787053606488398,
+ "loss": 5.466576099395752,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.018198711797595024,
+ "learning_rate": 0.0027863653299260356,
+ "loss": 5.447747230529785,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.015937626361846924,
+ "learning_rate": 0.00278567602904759,
+ "loss": 5.5055036544799805,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.01687259040772915,
+ "learning_rate": 0.002784985704408419,
+ "loss": 5.422250747680664,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.02622046135365963,
+ "learning_rate": 0.002784294356564705,
+ "loss": 5.466894626617432,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.018672186881303787,
+ "learning_rate": 0.0027836019860734535,
+ "loss": 5.48300838470459,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.02166241593658924,
+ "learning_rate": 0.0027829085934924958,
+ "loss": 5.477658271789551,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018426023423671722,
+ "learning_rate": 0.0027822141793804856,
+ "loss": 5.487186431884766,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.022229891270399094,
+ "learning_rate": 0.0027815187442968995,
+ "loss": 5.488579273223877,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.026930972933769226,
+ "learning_rate": 0.002780822288802037,
+ "loss": 5.492000102996826,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.02733265422284603,
+ "learning_rate": 0.00278012481345702,
+ "loss": 5.4663801193237305,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.02993253618478775,
+ "learning_rate": 0.0027794263188237922,
+ "loss": 5.543368816375732,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03152153640985489,
+ "learning_rate": 0.002778726805465118,
+ "loss": 5.565644264221191,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.031395044177770615,
+ "learning_rate": 0.002778026273944581,
+ "loss": 5.553359508514404,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.026603225618600845,
+ "learning_rate": 0.002777324724826589,
+ "loss": 5.552347183227539,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.024152139201760292,
+ "learning_rate": 0.002776622158676366,
+ "loss": 5.493257522583008,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.602636814117432,
+ "eval_runtime": 54.6199,
+ "eval_samples_per_second": 44.709,
+ "eval_steps_per_second": 1.41,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.02089761756360531,
+ "learning_rate": 0.0027759185760599576,
+ "loss": 5.461747646331787,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.018181778490543365,
+ "learning_rate": 0.0027752139775442278,
+ "loss": 5.492741584777832,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.020166238769888878,
+ "learning_rate": 0.0027745083636968575,
+ "loss": 5.476003646850586,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.015491694211959839,
+ "learning_rate": 0.0027738017350863488,
+ "loss": 5.425764560699463,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.022291697561740875,
+ "learning_rate": 0.0027730940922820182,
+ "loss": 5.45866584777832,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.018074683845043182,
+ "learning_rate": 0.002772385435854002,
+ "loss": 5.465143203735352,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.016872268170118332,
+ "learning_rate": 0.002771675766373251,
+ "loss": 5.446003437042236,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.023103339597582817,
+ "learning_rate": 0.0027709650844115336,
+ "loss": 5.411036491394043,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.027957158163189888,
+ "learning_rate": 0.0027702533905414324,
+ "loss": 5.452442169189453,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.019935399293899536,
+ "learning_rate": 0.002769540685336348,
+ "loss": 5.412145614624023,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.014351371675729752,
+ "learning_rate": 0.0027688269693704926,
+ "loss": 5.434524059295654,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.0152254244312644,
+ "learning_rate": 0.0027681122432188948,
+ "loss": 5.428544998168945,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.015071330592036247,
+ "learning_rate": 0.0027673965074573965,
+ "loss": 5.459929466247559,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.023415958508849144,
+ "learning_rate": 0.0027666797626626537,
+ "loss": 5.447159767150879,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.017228802666068077,
+ "learning_rate": 0.0027659620094121337,
+ "loss": 5.442273139953613,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.014184504747390747,
+ "learning_rate": 0.002765243248284118,
+ "loss": 5.448305130004883,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.024384526535868645,
+ "learning_rate": 0.0027645234798576988,
+ "loss": 5.4076337814331055,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.043152231723070145,
+ "learning_rate": 0.0027638027047127813,
+ "loss": 5.45344352722168,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.02889758162200451,
+ "learning_rate": 0.002763080923430081,
+ "loss": 5.489194393157959,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.02686968818306923,
+ "learning_rate": 0.002762358136591123,
+ "loss": 5.493659496307373,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.020467868074774742,
+ "learning_rate": 0.002761634344778245,
+ "loss": 5.410489559173584,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.019759224727749825,
+ "learning_rate": 0.002760909548574592,
+ "loss": 5.439385890960693,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.0233843382447958,
+ "learning_rate": 0.0027601837485641194,
+ "loss": 5.445036888122559,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.031753458082675934,
+ "learning_rate": 0.0027594569453315916,
+ "loss": 5.503895282745361,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.02089349366724491,
+ "learning_rate": 0.002758729139462581,
+ "loss": 5.45689582824707,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.03339357674121857,
+ "learning_rate": 0.002758000331543466,
+ "loss": 5.522361755371094,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019328996539115906,
+ "learning_rate": 0.002757270522161436,
+ "loss": 5.441373825073242,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019863856956362724,
+ "learning_rate": 0.002756539711904484,
+ "loss": 5.4633283615112305,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.02141754701733589,
+ "learning_rate": 0.0027558079013614116,
+ "loss": 5.453215599060059,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016591915860772133,
+ "learning_rate": 0.002755075091121824,
+ "loss": 5.463696479797363,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.01955028623342514,
+ "learning_rate": 0.0027543412817761345,
+ "loss": 5.411355018615723,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.016791662201285362,
+ "learning_rate": 0.0027536064739155596,
+ "loss": 5.459614276885986,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.019661981612443924,
+ "learning_rate": 0.00275287066813212,
+ "loss": 5.440699577331543,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.01927965320646763,
+ "learning_rate": 0.002752133865018642,
+ "loss": 5.409865856170654,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.016483580693602562,
+ "learning_rate": 0.002751396065168754,
+ "loss": 5.390999794006348,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016264773905277252,
+ "learning_rate": 0.0027506572691768888,
+ "loss": 5.418543815612793,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.01958460919559002,
+ "learning_rate": 0.00274991747763828,
+ "loss": 5.403289794921875,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.01861695572733879,
+ "learning_rate": 0.002749176691148965,
+ "loss": 5.431911468505859,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.029594263061881065,
+ "learning_rate": 0.002748434910305782,
+ "loss": 5.399242401123047,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.028574613854289055,
+ "learning_rate": 0.0027476921357063697,
+ "loss": 5.440446853637695,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.028612585738301277,
+ "learning_rate": 0.0027469483679491693,
+ "loss": 5.410346031188965,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.020704898983240128,
+ "learning_rate": 0.00274620360763342,
+ "loss": 5.44694709777832,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.020542364567518234,
+ "learning_rate": 0.002745457855359163,
+ "loss": 5.402653694152832,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.015762662515044212,
+ "learning_rate": 0.0027447111117272363,
+ "loss": 5.396594047546387,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022096984088420868,
+ "learning_rate": 0.002743963377339278,
+ "loss": 5.398744106292725,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.018791764974594116,
+ "learning_rate": 0.002743214652797724,
+ "loss": 5.397758483886719,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01642526686191559,
+ "learning_rate": 0.0027424649387058093,
+ "loss": 5.411459922790527,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.017931703478097916,
+ "learning_rate": 0.0027417142356675637,
+ "loss": 5.390264511108398,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018819602206349373,
+ "learning_rate": 0.0027409625442878153,
+ "loss": 5.407539367675781,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.019389411434531212,
+ "learning_rate": 0.0027402098651721886,
+ "loss": 5.386474609375,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.013645983301103115,
+ "learning_rate": 0.002739456198927104,
+ "loss": 5.373845100402832,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.013846895657479763,
+ "learning_rate": 0.002738701546159776,
+ "loss": 5.387155055999756,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013553125783801079,
+ "learning_rate": 0.002737945907478215,
+ "loss": 5.377801895141602,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.012771612964570522,
+ "learning_rate": 0.002737189283491225,
+ "loss": 5.344066143035889,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.020465301349759102,
+ "learning_rate": 0.002736431674808405,
+ "loss": 5.382771968841553,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.014798106625676155,
+ "learning_rate": 0.0027356730820401465,
+ "loss": 5.370522499084473,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.01566634140908718,
+ "learning_rate": 0.0027349135057976332,
+ "loss": 5.351980209350586,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.017485102638602257,
+ "learning_rate": 0.002734152946692843,
+ "loss": 5.355443477630615,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.016646096482872963,
+ "learning_rate": 0.0027333914053385444,
+ "loss": 5.34829044342041,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.017293352633714676,
+ "learning_rate": 0.0027326288823482964,
+ "loss": 5.3519487380981445,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.018873002380132675,
+ "learning_rate": 0.002731865378336452,
+ "loss": 5.348442077636719,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.015504784882068634,
+ "learning_rate": 0.002731100893918151,
+ "loss": 5.341490268707275,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.014270870015025139,
+ "learning_rate": 0.0027303354297093254,
+ "loss": 5.359251022338867,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.018820680677890778,
+ "learning_rate": 0.002729568986326695,
+ "loss": 5.390280723571777,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.021008187904953957,
+ "learning_rate": 0.0027288015643877708,
+ "loss": 5.338380336761475,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.016266655176877975,
+ "learning_rate": 0.0027280331645108498,
+ "loss": 5.347043037414551,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.01831045374274254,
+ "learning_rate": 0.0027272637873150185,
+ "loss": 5.326537132263184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02073794975876808,
+ "learning_rate": 0.002726493433420149,
+ "loss": 5.3607025146484375,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.022534754127264023,
+ "learning_rate": 0.0027257221034469033,
+ "loss": 5.353292465209961,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.021627362817525864,
+ "learning_rate": 0.002724949798016727,
+ "loss": 5.373323917388916,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.018226243555545807,
+ "learning_rate": 0.0027241765177518523,
+ "loss": 5.344578266143799,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.02032875455915928,
+ "learning_rate": 0.002723402263275298,
+ "loss": 5.329159259796143,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.44969367980957,
+ "eval_runtime": 51.2294,
+ "eval_samples_per_second": 47.668,
+ "eval_steps_per_second": 1.503,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.01613701693713665,
+ "learning_rate": 0.0027226270352108664,
+ "loss": 5.3024678230285645,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01940973661839962,
+ "learning_rate": 0.0027218508341831445,
+ "loss": 5.270181179046631,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.015840493142604828,
+ "learning_rate": 0.002721073660817504,
+ "loss": 5.261706352233887,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.017390675842761993,
+ "learning_rate": 0.0027202955157401,
+ "loss": 5.262207984924316,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.015184012241661549,
+ "learning_rate": 0.0027195163995778686,
+ "loss": 5.297155380249023,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.013417073525488377,
+ "learning_rate": 0.00271873631295853,
+ "loss": 5.279838562011719,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.013024317100644112,
+ "learning_rate": 0.0027179552565105864,
+ "loss": 5.265446186065674,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.01426338218152523,
+ "learning_rate": 0.002717173230863321,
+ "loss": 5.284180164337158,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.019604040309786797,
+ "learning_rate": 0.002716390236646797,
+ "loss": 5.286281108856201,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.019039476290345192,
+ "learning_rate": 0.0027156062744918587,
+ "loss": 5.290433406829834,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.019284334033727646,
+ "learning_rate": 0.0027148213450301313,
+ "loss": 5.262433052062988,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.0145154669880867,
+ "learning_rate": 0.0027140354488940175,
+ "loss": 5.24387264251709,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.014776417054235935,
+ "learning_rate": 0.0027132485867166992,
+ "loss": 5.282435417175293,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.012711729854345322,
+ "learning_rate": 0.0027124607591321377,
+ "loss": 5.246735572814941,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.013394835405051708,
+ "learning_rate": 0.0027116719667750716,
+ "loss": 5.270739555358887,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.013468850404024124,
+ "learning_rate": 0.002710882210281016,
+ "loss": 5.27161169052124,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.030262840911746025,
+ "learning_rate": 0.002710091490286264,
+ "loss": 5.256673812866211,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.013560288585722446,
+ "learning_rate": 0.0027092998074278846,
+ "loss": 5.253674507141113,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014592813327908516,
+ "learning_rate": 0.002708507162343721,
+ "loss": 5.25804328918457,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014328169636428356,
+ "learning_rate": 0.0027077135556723953,
+ "loss": 5.235117435455322,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.01277366653084755,
+ "learning_rate": 0.0027069189880533,
+ "loss": 5.2053985595703125,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.012948785908520222,
+ "learning_rate": 0.0027061234601266043,
+ "loss": 5.214653015136719,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.014892828650772572,
+ "learning_rate": 0.002705326972533252,
+ "loss": 5.227547645568848,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01427230704575777,
+ "learning_rate": 0.002704529525914957,
+ "loss": 5.241065502166748,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.01650116965174675,
+ "learning_rate": 0.0027037311209142092,
+ "loss": 5.227713108062744,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.019728578627109528,
+ "learning_rate": 0.0027029317581742677,
+ "loss": 5.2624921798706055,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.023847149685025215,
+ "learning_rate": 0.0027021314383391657,
+ "loss": 5.278811454772949,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.020690469071269035,
+ "learning_rate": 0.0027013301620537056,
+ "loss": 5.238919258117676,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.021211959421634674,
+ "learning_rate": 0.0027005279299634616,
+ "loss": 5.253684997558594,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.025854922831058502,
+ "learning_rate": 0.002699724742714778,
+ "loss": 5.2538652420043945,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.0215485580265522,
+ "learning_rate": 0.0026989206009547678,
+ "loss": 5.224540710449219,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.019827792420983315,
+ "learning_rate": 0.0026981155053313137,
+ "loss": 5.220674514770508,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.017689093947410583,
+ "learning_rate": 0.002697309456493066,
+ "loss": 5.228511810302734,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.022564806044101715,
+ "learning_rate": 0.002696502455089445,
+ "loss": 5.295933723449707,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.020070280879735947,
+ "learning_rate": 0.0026956945017706357,
+ "loss": 5.230920791625977,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.025744661688804626,
+ "learning_rate": 0.002694885597187593,
+ "loss": 5.296453952789307,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.021263007074594498,
+ "learning_rate": 0.0026940757419920357,
+ "loss": 5.220481872558594,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.017577245831489563,
+ "learning_rate": 0.0026932649368364495,
+ "loss": 5.231488227844238,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.015562832355499268,
+ "learning_rate": 0.002692453182374086,
+ "loss": 5.175224304199219,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.014115403406322002,
+ "learning_rate": 0.0026916404792589606,
+ "loss": 5.228387832641602,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.012863830663263798,
+ "learning_rate": 0.002690826828145854,
+ "loss": 5.186727523803711,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.014151126146316528,
+ "learning_rate": 0.002690012229690309,
+ "loss": 5.1694817543029785,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.01316932775080204,
+ "learning_rate": 0.0026891966845486353,
+ "loss": 5.19931697845459,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.011340726166963577,
+ "learning_rate": 0.0026883801933779003,
+ "loss": 5.238619804382324,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.012616089545190334,
+ "learning_rate": 0.0026875627568359378,
+ "loss": 5.183376789093018,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.012274038046598434,
+ "learning_rate": 0.002686744375581341,
+ "loss": 5.199946403503418,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.014990398660302162,
+ "learning_rate": 0.002685925050273465,
+ "loss": 5.168451309204102,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.011399757117033005,
+ "learning_rate": 0.0026851047815724255,
+ "loss": 5.173828125,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.013638390228152275,
+ "learning_rate": 0.0026842835701390982,
+ "loss": 5.142702102661133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.014472718350589275,
+ "learning_rate": 0.0026834614166351177,
+ "loss": 5.173436164855957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.014509808272123337,
+ "learning_rate": 0.002682638321722879,
+ "loss": 5.173975467681885,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.013576583005487919,
+ "learning_rate": 0.0026818142860655352,
+ "loss": 5.1435089111328125,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.012914472259581089,
+ "learning_rate": 0.0026809893103269954,
+ "loss": 5.151381492614746,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.013803352601826191,
+ "learning_rate": 0.0026801633951719288,
+ "loss": 5.133159637451172,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.018488213419914246,
+ "learning_rate": 0.0026793365412657606,
+ "loss": 5.183903694152832,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.01843969337642193,
+ "learning_rate": 0.002678508749274671,
+ "loss": 5.1660356521606445,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.017181839793920517,
+ "learning_rate": 0.0026776800198655983,
+ "loss": 5.179847717285156,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0148160420358181,
+ "learning_rate": 0.0026768503537062333,
+ "loss": 5.149198532104492,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.013251296244561672,
+ "learning_rate": 0.0026760197514650247,
+ "loss": 5.192000389099121,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.010746927931904793,
+ "learning_rate": 0.0026751882138111717,
+ "loss": 5.138680458068848,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.012316994369029999,
+ "learning_rate": 0.002674355741414631,
+ "loss": 5.169965744018555,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.01191797386854887,
+ "learning_rate": 0.0026735223349461106,
+ "loss": 5.128267288208008,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.013525084592401981,
+ "learning_rate": 0.0026726879950770698,
+ "loss": 5.124795436859131,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.014843370765447617,
+ "learning_rate": 0.002671852722479721,
+ "loss": 5.1497087478637695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.015285412780940533,
+ "learning_rate": 0.00267101651782703,
+ "loss": 5.093118190765381,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.013555832207202911,
+ "learning_rate": 0.0026701793817927107,
+ "loss": 5.13249397277832,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.010569742880761623,
+ "learning_rate": 0.0026693413150512273,
+ "loss": 5.084077835083008,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.011483813636004925,
+ "learning_rate": 0.0026685023182777975,
+ "loss": 5.128049850463867,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.013724366202950478,
+ "learning_rate": 0.0026676623921483844,
+ "loss": 5.157539367675781,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.014337432570755482,
+ "learning_rate": 0.0026668215373397016,
+ "loss": 5.149948596954346,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.014101454056799412,
+ "learning_rate": 0.00266597975452921,
+ "loss": 5.117250442504883,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.016004368662834167,
+ "learning_rate": 0.0026651370443951197,
+ "loss": 5.143221855163574,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.249833583831787,
+ "eval_runtime": 51.5543,
+ "eval_samples_per_second": 47.368,
+ "eval_steps_per_second": 1.494,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.01297403872013092,
+ "learning_rate": 0.0026642934076163867,
+ "loss": 5.089287281036377,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.012867718935012817,
+ "learning_rate": 0.0026634488448727144,
+ "loss": 5.0784101486206055,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.014600927010178566,
+ "learning_rate": 0.0026626033568445504,
+ "loss": 5.050220489501953,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.016619572415947914,
+ "learning_rate": 0.0026617569442130905,
+ "loss": 5.079303741455078,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.015560587868094444,
+ "learning_rate": 0.0026609096076602736,
+ "loss": 5.072385311126709,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.01774057187139988,
+ "learning_rate": 0.0026600613478687833,
+ "loss": 5.122529983520508,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.019971400499343872,
+ "learning_rate": 0.0026592121655220472,
+ "loss": 5.085966110229492,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.017667256295681,
+ "learning_rate": 0.0026583620613042365,
+ "loss": 5.078432559967041,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.022369660437107086,
+ "learning_rate": 0.002657511035900264,
+ "loss": 5.0630903244018555,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.017964458093047142,
+ "learning_rate": 0.002656659089995787,
+ "loss": 5.0846076011657715,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.021769145503640175,
+ "learning_rate": 0.002655806224277201,
+ "loss": 5.081344127655029,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.02487620897591114,
+ "learning_rate": 0.0026549524394316464,
+ "loss": 5.106287956237793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.022516218945384026,
+ "learning_rate": 0.0026540977361470012,
+ "loss": 5.128098487854004,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.02099926583468914,
+ "learning_rate": 0.0026532421151118845,
+ "loss": 5.103562355041504,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.01899387687444687,
+ "learning_rate": 0.002652385577015655,
+ "loss": 5.062675476074219,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.016801996156573296,
+ "learning_rate": 0.002651528122548409,
+ "loss": 5.096002578735352,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.014009200967848301,
+ "learning_rate": 0.0026506697524009836,
+ "loss": 5.075201034545898,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.014653858728706837,
+ "learning_rate": 0.002649810467264951,
+ "loss": 5.090478897094727,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.014918150380253792,
+ "learning_rate": 0.0026489502678326226,
+ "loss": 5.070797920227051,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.01327288057655096,
+ "learning_rate": 0.0026480891547970443,
+ "loss": 5.030500411987305,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.012885140255093575,
+ "learning_rate": 0.002647227128852,
+ "loss": 5.0845537185668945,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.012848512269556522,
+ "learning_rate": 0.002646364190692009,
+ "loss": 5.079484939575195,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.01149789709597826,
+ "learning_rate": 0.002645500341012324,
+ "loss": 5.036316394805908,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.010360757820308208,
+ "learning_rate": 0.0026446355805089337,
+ "loss": 5.04129695892334,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.010426290333271027,
+ "learning_rate": 0.00264376990987856,
+ "loss": 5.051270484924316,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.01018153689801693,
+ "learning_rate": 0.002642903329818658,
+ "loss": 5.102342128753662,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.011879283003509045,
+ "learning_rate": 0.0026420358410274153,
+ "loss": 5.053501129150391,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.010418329387903214,
+ "learning_rate": 0.002641167444203752,
+ "loss": 5.030653476715088,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.012302013114094734,
+ "learning_rate": 0.0026402981400473202,
+ "loss": 5.0402069091796875,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.011392916552722454,
+ "learning_rate": 0.002639427929258502,
+ "loss": 5.024111747741699,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.013075058348476887,
+ "learning_rate": 0.0026385568125384114,
+ "loss": 5.021731376647949,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015349832363426685,
+ "learning_rate": 0.0026376847905888903,
+ "loss": 5.03192138671875,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01588309183716774,
+ "learning_rate": 0.0026368118641125113,
+ "loss": 4.966363906860352,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.014378590509295464,
+ "learning_rate": 0.0026359380338125764,
+ "loss": 5.070773124694824,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.012306655757129192,
+ "learning_rate": 0.002635063300393114,
+ "loss": 5.025704860687256,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01212004292756319,
+ "learning_rate": 0.002634187664558881,
+ "loss": 4.965104103088379,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013538849540054798,
+ "learning_rate": 0.0026333111270153624,
+ "loss": 5.029810905456543,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.012052906677126884,
+ "learning_rate": 0.0026324336884687672,
+ "loss": 5.0262908935546875,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.011368660256266594,
+ "learning_rate": 0.002631555349626033,
+ "loss": 5.001485824584961,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.010215395130217075,
+ "learning_rate": 0.0026306761111948205,
+ "loss": 5.019739151000977,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.011449878104031086,
+ "learning_rate": 0.0026297959738835167,
+ "loss": 4.9816083908081055,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.012381811626255512,
+ "learning_rate": 0.0026289149384012326,
+ "loss": 5.025032043457031,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.00974641740322113,
+ "learning_rate": 0.002628033005457802,
+ "loss": 5.000177383422852,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.009515874087810516,
+ "learning_rate": 0.0026271501757637833,
+ "loss": 4.996979713439941,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.011125455610454082,
+ "learning_rate": 0.0026262664500304555,
+ "loss": 4.996796607971191,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.013290468603372574,
+ "learning_rate": 0.0026253818289698207,
+ "loss": 5.053703308105469,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.014650283381342888,
+ "learning_rate": 0.002624496313294602,
+ "loss": 5.001321792602539,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014537827111780643,
+ "learning_rate": 0.002623609903718244,
+ "loss": 5.024584770202637,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.012908566743135452,
+ "learning_rate": 0.0026227226009549105,
+ "loss": 5.042878150939941,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.01446808222681284,
+ "learning_rate": 0.002621834405719485,
+ "loss": 5.020857810974121,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.01535153016448021,
+ "learning_rate": 0.0026209453187275708,
+ "loss": 5.0056304931640625,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01815616339445114,
+ "learning_rate": 0.0026200553406954892,
+ "loss": 5.0467376708984375,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.023135516792535782,
+ "learning_rate": 0.0026191644723402795,
+ "loss": 5.033995628356934,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.019682930782437325,
+ "learning_rate": 0.002618272714379698,
+ "loss": 5.008149147033691,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.018521904945373535,
+ "learning_rate": 0.002617380067532217,
+ "loss": 5.007076263427734,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.02118663489818573,
+ "learning_rate": 0.0026164865325170278,
+ "loss": 5.05519962310791,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024684788659214973,
+ "learning_rate": 0.0026155921100540344,
+ "loss": 5.027571201324463,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.02609819918870926,
+ "learning_rate": 0.0026146968008638567,
+ "loss": 5.10361385345459,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.025958053767681122,
+ "learning_rate": 0.00261380060566783,
+ "loss": 5.0911455154418945,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.017067518085241318,
+ "learning_rate": 0.002612903525188002,
+ "loss": 5.045783042907715,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.016510486602783203,
+ "learning_rate": 0.0026120055601471345,
+ "loss": 5.069464683532715,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015552333556115627,
+ "learning_rate": 0.0026111067112687012,
+ "loss": 5.024248123168945,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.013432685285806656,
+ "learning_rate": 0.0026102069792768886,
+ "loss": 5.004837989807129,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.012300201691687107,
+ "learning_rate": 0.0026093063648965943,
+ "loss": 5.031093597412109,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.01197744719684124,
+ "learning_rate": 0.002608404868853428,
+ "loss": 4.965237617492676,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.01100246049463749,
+ "learning_rate": 0.002607502491873708,
+ "loss": 5.04719352722168,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.010111447423696518,
+ "learning_rate": 0.0026065992346844626,
+ "loss": 4.986410617828369,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.010669790208339691,
+ "learning_rate": 0.002605695098013431,
+ "loss": 4.9825663566589355,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.010805477388203144,
+ "learning_rate": 0.0026047900825890593,
+ "loss": 4.991506576538086,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.010886074975132942,
+ "learning_rate": 0.002603884189140502,
+ "loss": 4.987531661987305,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.010735839605331421,
+ "learning_rate": 0.0026029774183976206,
+ "loss": 4.982756614685059,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.011092593893408775,
+ "learning_rate": 0.002602069771090985,
+ "loss": 4.994824409484863,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 5.13364315032959,
+ "eval_runtime": 50.4195,
+ "eval_samples_per_second": 48.434,
+ "eval_steps_per_second": 1.527,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.01005905494093895,
+ "learning_rate": 0.0026011612479518694,
+ "loss": 4.918412685394287,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.009904682636260986,
+ "learning_rate": 0.002600251849712255,
+ "loss": 4.924630165100098,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.009543103165924549,
+ "learning_rate": 0.0025993415771048273,
+ "loss": 4.940309524536133,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.011031470261514187,
+ "learning_rate": 0.0025984304308629765,
+ "loss": 4.924609184265137,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.009511022828519344,
+ "learning_rate": 0.002597518411720796,
+ "loss": 4.9102325439453125,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.00886621605604887,
+ "learning_rate": 0.0025966055204130847,
+ "loss": 4.9236907958984375,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.010057039558887482,
+ "learning_rate": 0.002595691757675341,
+ "loss": 4.908633232116699,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.01098555512726307,
+ "learning_rate": 0.0025947771242437684,
+ "loss": 4.935927391052246,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.008800056762993336,
+ "learning_rate": 0.0025938616208552694,
+ "loss": 4.873291969299316,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.009747766889631748,
+ "learning_rate": 0.0025929452482474497,
+ "loss": 4.954436302185059,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.010785684920847416,
+ "learning_rate": 0.002592028007158613,
+ "loss": 4.924927711486816,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.012620717287063599,
+ "learning_rate": 0.0025911098983277645,
+ "loss": 4.953606605529785,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.014428738504648209,
+ "learning_rate": 0.002590190922494607,
+ "loss": 4.951706886291504,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.014496048912405968,
+ "learning_rate": 0.0025892710803995444,
+ "loss": 4.941249847412109,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.0109238987788558,
+ "learning_rate": 0.0025883503727836764,
+ "loss": 4.939975261688232,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.012271072715520859,
+ "learning_rate": 0.0025874288003887996,
+ "loss": 4.880942344665527,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.016302887350320816,
+ "learning_rate": 0.0025865063639574086,
+ "loss": 4.931602478027344,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.016137251630425453,
+ "learning_rate": 0.0025855830642326945,
+ "loss": 4.918013095855713,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.014732961542904377,
+ "learning_rate": 0.002584658901958542,
+ "loss": 4.927577495574951,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.016387946903705597,
+ "learning_rate": 0.0025837338778795325,
+ "loss": 4.930819034576416,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.017146151512861252,
+ "learning_rate": 0.002582807992740941,
+ "loss": 4.928377151489258,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.016153613105416298,
+ "learning_rate": 0.0025818812472887375,
+ "loss": 4.975607395172119,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.013386047445237637,
+ "learning_rate": 0.0025809536422695825,
+ "loss": 4.9434709548950195,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.012328253127634525,
+ "learning_rate": 0.0025800251784308314,
+ "loss": 4.917585372924805,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.012000586837530136,
+ "learning_rate": 0.0025790958565205316,
+ "loss": 4.9137725830078125,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.009463997557759285,
+ "learning_rate": 0.002578165677287419,
+ "loss": 4.910362243652344,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.010591360740363598,
+ "learning_rate": 0.0025772346414809242,
+ "loss": 4.916032314300537,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.010429762303829193,
+ "learning_rate": 0.002576302749851164,
+ "loss": 4.934561729431152,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.12526583671569824,
+ "learning_rate": 0.0025753700031489486,
+ "loss": 4.897515773773193,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01368892751634121,
+ "learning_rate": 0.0025744364021257743,
+ "loss": 4.94814920425415,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.016603967174887657,
+ "learning_rate": 0.0025735019475338254,
+ "loss": 4.8996758460998535,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.06110106781125069,
+ "learning_rate": 0.0025725666401259775,
+ "loss": 4.946794509887695,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.018292423337697983,
+ "learning_rate": 0.002571630480655789,
+ "loss": 4.958992958068848,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.01852245442569256,
+ "learning_rate": 0.0025706934698775073,
+ "loss": 4.942556381225586,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.0195649154484272,
+ "learning_rate": 0.0025697556085460646,
+ "loss": 4.895773887634277,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.017820244655013084,
+ "learning_rate": 0.00256881689741708,
+ "loss": 4.915250778198242,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.016814742237329483,
+ "learning_rate": 0.002567877337246855,
+ "loss": 4.9521260261535645,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.015319781377911568,
+ "learning_rate": 0.002566936928792377,
+ "loss": 4.899402141571045,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.01607401855289936,
+ "learning_rate": 0.0025659956728113146,
+ "loss": 4.909913063049316,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.015564955770969391,
+ "learning_rate": 0.002565053570062023,
+ "loss": 4.947719573974609,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.017502037808299065,
+ "learning_rate": 0.002564110621303536,
+ "loss": 4.95107364654541,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.014791073277592659,
+ "learning_rate": 0.002563166827295571,
+ "loss": 4.8995866775512695,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.012700124643743038,
+ "learning_rate": 0.002562222188798525,
+ "loss": 4.922100067138672,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.011235272511839867,
+ "learning_rate": 0.0025612767065734763,
+ "loss": 4.884456634521484,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.013009768910706043,
+ "learning_rate": 0.002560330381382183,
+ "loss": 4.869503021240234,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.0102638378739357,
+ "learning_rate": 0.0025593832139870834,
+ "loss": 4.934551239013672,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.011081870645284653,
+ "learning_rate": 0.0025584352051512913,
+ "loss": 4.937268257141113,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.010681218467652798,
+ "learning_rate": 0.002557486355638601,
+ "loss": 4.920559883117676,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.010962890461087227,
+ "learning_rate": 0.002556536666213484,
+ "loss": 4.936082363128662,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.010790379717946053,
+ "learning_rate": 0.0025555861376410868,
+ "loss": 4.9506449699401855,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.010659157298505306,
+ "learning_rate": 0.0025546347706872346,
+ "loss": 4.872883319854736,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.012962655164301395,
+ "learning_rate": 0.0025536825661184254,
+ "loss": 4.901210784912109,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.014810293912887573,
+ "learning_rate": 0.002552729524701833,
+ "loss": 4.895804405212402,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.010993832722306252,
+ "learning_rate": 0.0025517756472053068,
+ "loss": 4.891035079956055,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.013413915410637856,
+ "learning_rate": 0.002550820934397368,
+ "loss": 4.900317192077637,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.01580147072672844,
+ "learning_rate": 0.0025498653870472117,
+ "loss": 4.883669853210449,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01864512264728546,
+ "learning_rate": 0.0025489090059247045,
+ "loss": 4.881938457489014,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017185743898153305,
+ "learning_rate": 0.0025479517918003862,
+ "loss": 4.853325843811035,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.012045156210660934,
+ "learning_rate": 0.002546993745445466,
+ "loss": 4.9152326583862305,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.013420728035271168,
+ "learning_rate": 0.0025460348676318257,
+ "loss": 4.895130157470703,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.011561610735952854,
+ "learning_rate": 0.0025450751591320152,
+ "loss": 4.890257835388184,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013164008036255836,
+ "learning_rate": 0.0025441146207192547,
+ "loss": 4.885972023010254,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01120204571634531,
+ "learning_rate": 0.002543153253167432,
+ "loss": 4.882246971130371,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.011342627927660942,
+ "learning_rate": 0.002542191057251104,
+ "loss": 4.862927436828613,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.010852213948965073,
+ "learning_rate": 0.0025412280337454942,
+ "loss": 4.879744529724121,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.012851282022893429,
+ "learning_rate": 0.0025402641834264936,
+ "loss": 4.882238388061523,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.011839655227959156,
+ "learning_rate": 0.002539299507070659,
+ "loss": 4.85658073425293,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.011315262876451015,
+ "learning_rate": 0.002538334005455213,
+ "loss": 4.9234771728515625,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.011661105789244175,
+ "learning_rate": 0.0025373676793580416,
+ "loss": 4.855696678161621,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.009712521918118,
+ "learning_rate": 0.0025364005295576975,
+ "loss": 4.89942741394043,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.011002928018569946,
+ "learning_rate": 0.0025354325568333956,
+ "loss": 4.871459007263184,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.01509513333439827,
+ "learning_rate": 0.0025344637619650135,
+ "loss": 4.834778785705566,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 5.031810283660889,
+ "eval_runtime": 51.9412,
+ "eval_samples_per_second": 47.015,
+ "eval_steps_per_second": 1.482,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.010317740961909294,
+ "learning_rate": 0.0025334941457330925,
+ "loss": 4.8169708251953125,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.00892715621739626,
+ "learning_rate": 0.0025325237089188343,
+ "loss": 4.852178573608398,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.011608307249844074,
+ "learning_rate": 0.002531552452304103,
+ "loss": 4.790487289428711,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.013100114651024342,
+ "learning_rate": 0.0025305803766714216,
+ "loss": 4.863536834716797,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.016571057960391045,
+ "learning_rate": 0.002529607482803974,
+ "loss": 4.840987205505371,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.014595622196793556,
+ "learning_rate": 0.002528633771485604,
+ "loss": 4.815027236938477,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.014603931456804276,
+ "learning_rate": 0.002527659243500812,
+ "loss": 4.853024959564209,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.019107041880488396,
+ "learning_rate": 0.0025266838996347595,
+ "loss": 4.829808235168457,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.020292190834879875,
+ "learning_rate": 0.0025257077406732613,
+ "loss": 4.775232315063477,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.014630074612796307,
+ "learning_rate": 0.0025247307674027926,
+ "loss": 4.826682090759277,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.019507668912410736,
+ "learning_rate": 0.0025237529806104823,
+ "loss": 4.843222141265869,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.023950330913066864,
+ "learning_rate": 0.002522774381084116,
+ "loss": 4.820512771606445,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.022547820582985878,
+ "learning_rate": 0.002521794969612133,
+ "loss": 4.854910850524902,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.017819277942180634,
+ "learning_rate": 0.0025208147469836283,
+ "loss": 4.825885772705078,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.017989592626690865,
+ "learning_rate": 0.002519833713988349,
+ "loss": 4.8357014656066895,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.013516640290617943,
+ "learning_rate": 0.0025188518714166956,
+ "loss": 4.83729362487793,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.015956101939082146,
+ "learning_rate": 0.002517869220059721,
+ "loss": 4.807845592498779,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.013967473991215229,
+ "learning_rate": 0.002516885760709131,
+ "loss": 4.828944206237793,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.01270931027829647,
+ "learning_rate": 0.0025159014941572785,
+ "loss": 4.826594352722168,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.012433716095983982,
+ "learning_rate": 0.002514916421197171,
+ "loss": 4.797939777374268,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.011882808059453964,
+ "learning_rate": 0.0025139305426224636,
+ "loss": 4.78917932510376,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.011193005368113518,
+ "learning_rate": 0.0025129438592274614,
+ "loss": 4.787964820861816,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.011873407289385796,
+ "learning_rate": 0.002511956371807116,
+ "loss": 4.81754732131958,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.011407418176531792,
+ "learning_rate": 0.002510968081157029,
+ "loss": 4.781639575958252,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.009560304693877697,
+ "learning_rate": 0.002509978988073449,
+ "loss": 4.8163957595825195,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.009969786740839481,
+ "learning_rate": 0.0025089890933532697,
+ "loss": 4.795045852661133,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.010739767923951149,
+ "learning_rate": 0.0025079983977940312,
+ "loss": 4.847584247589111,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.009999158792197704,
+ "learning_rate": 0.0025070069021939194,
+ "loss": 4.834909439086914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.009302482940256596,
+ "learning_rate": 0.002506014607351765,
+ "loss": 4.808745384216309,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.010687744244933128,
+ "learning_rate": 0.0025050215140670403,
+ "loss": 4.789101600646973,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.010930616408586502,
+ "learning_rate": 0.0025040276231398637,
+ "loss": 4.817464351654053,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.013101065531373024,
+ "learning_rate": 0.0025030329353709956,
+ "loss": 4.827780723571777,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.01223307941108942,
+ "learning_rate": 0.0025020374515618373,
+ "loss": 4.8367414474487305,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.013091634958982468,
+ "learning_rate": 0.0025010411725144323,
+ "loss": 4.839454650878906,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.012114843353629112,
+ "learning_rate": 0.0025000440990314643,
+ "loss": 4.832447052001953,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014664588496088982,
+ "learning_rate": 0.0024990462319162588,
+ "loss": 4.740812301635742,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.015566283836960793,
+ "learning_rate": 0.002498047571972777,
+ "loss": 4.815671920776367,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.012440434657037258,
+ "learning_rate": 0.002497048120005623,
+ "loss": 4.769775867462158,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01175499614328146,
+ "learning_rate": 0.0024960478768200362,
+ "loss": 4.8101348876953125,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.012699508108198643,
+ "learning_rate": 0.002495046843221896,
+ "loss": 4.827281951904297,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.012438303790986538,
+ "learning_rate": 0.002494045020017715,
+ "loss": 4.786289691925049,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.01709924265742302,
+ "learning_rate": 0.002493042408014646,
+ "loss": 4.828742027282715,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.022684535011649132,
+ "learning_rate": 0.0024920390080204733,
+ "loss": 4.822258949279785,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.011984334327280521,
+ "learning_rate": 0.00249103482084362,
+ "loss": 4.792006969451904,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.013941102661192417,
+ "learning_rate": 0.00249002984729314,
+ "loss": 4.7926177978515625,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.0145634226500988,
+ "learning_rate": 0.002489024088178724,
+ "loss": 4.819155216217041,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.013391612097620964,
+ "learning_rate": 0.0024880175443106932,
+ "loss": 4.777242660522461,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01764756441116333,
+ "learning_rate": 0.0024870102165000015,
+ "loss": 4.81627893447876,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.015397441573441029,
+ "learning_rate": 0.002486002105558235,
+ "loss": 4.869495391845703,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01225620973855257,
+ "learning_rate": 0.0024849932122976105,
+ "loss": 4.784395217895508,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.01534759160131216,
+ "learning_rate": 0.002483983537530975,
+ "loss": 4.813475608825684,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.015057779848575592,
+ "learning_rate": 0.002482973082071805,
+ "loss": 4.8281569480896,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.01243986189365387,
+ "learning_rate": 0.0024819618467342066,
+ "loss": 4.7866973876953125,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.013326527550816536,
+ "learning_rate": 0.0024809498323329136,
+ "loss": 4.750214576721191,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.013644695281982422,
+ "learning_rate": 0.002479937039683288,
+ "loss": 4.790144920349121,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01296213548630476,
+ "learning_rate": 0.002478923469601319,
+ "loss": 4.812170028686523,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.011848201975226402,
+ "learning_rate": 0.0024779091229036204,
+ "loss": 4.798855781555176,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.012765968218445778,
+ "learning_rate": 0.0024768940004074335,
+ "loss": 4.8272294998168945,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.011700666509568691,
+ "learning_rate": 0.002475878102930625,
+ "loss": 4.762265205383301,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.01018498931080103,
+ "learning_rate": 0.0024748614312916843,
+ "loss": 4.796628952026367,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.010241681709885597,
+ "learning_rate": 0.002473843986309725,
+ "loss": 4.829319000244141,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.009332080371677876,
+ "learning_rate": 0.0024728257688044855,
+ "loss": 4.783773899078369,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.012860726565122604,
+ "learning_rate": 0.002471806779596324,
+ "loss": 4.802700519561768,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.011156697757542133,
+ "learning_rate": 0.0024707870195062225,
+ "loss": 4.808038234710693,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.010329720564186573,
+ "learning_rate": 0.002469766489355783,
+ "loss": 4.761645317077637,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.011093895882368088,
+ "learning_rate": 0.0024687451899672278,
+ "loss": 4.795220851898193,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.010740782134234905,
+ "learning_rate": 0.0024677231221634,
+ "loss": 4.79289436340332,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.012272508814930916,
+ "learning_rate": 0.0024667002867677606,
+ "loss": 4.811873435974121,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.008665879257023335,
+ "learning_rate": 0.0024656766846043905,
+ "loss": 4.744179725646973,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.00933373998850584,
+ "learning_rate": 0.0024646523164979866,
+ "loss": 4.805453777313232,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.010399337857961655,
+ "learning_rate": 0.002463627183273864,
+ "loss": 4.75568962097168,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.013210986740887165,
+ "learning_rate": 0.0024626012857579542,
+ "loss": 4.797974586486816,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.967026233673096,
+ "eval_runtime": 51.07,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.01731858029961586,
+ "learning_rate": 0.0024615746247768043,
+ "loss": 4.748838901519775,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.018902689218521118,
+ "learning_rate": 0.0024605472011575763,
+ "loss": 4.717338562011719,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.01734415628015995,
+ "learning_rate": 0.0024595190157280468,
+ "loss": 4.737737655639648,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.01971321739256382,
+ "learning_rate": 0.002458490069316607,
+ "loss": 4.774558067321777,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.020029844716191292,
+ "learning_rate": 0.002457460362752259,
+ "loss": 4.739579200744629,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.020853109657764435,
+ "learning_rate": 0.0024564298968646203,
+ "loss": 4.791436672210693,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.019159914925694466,
+ "learning_rate": 0.0024553986724839176,
+ "loss": 4.765683650970459,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.0165473110973835,
+ "learning_rate": 0.0024543666904409903,
+ "loss": 4.732603073120117,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.017426826059818268,
+ "learning_rate": 0.002453333951567287,
+ "loss": 4.7360639572143555,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014590734615921974,
+ "learning_rate": 0.002452300456694868,
+ "loss": 4.750031471252441,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.011172222904860973,
+ "learning_rate": 0.0024512662066563996,
+ "loss": 4.740771293640137,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.012294141575694084,
+ "learning_rate": 0.0024502312022851606,
+ "loss": 4.726214408874512,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.01146282535046339,
+ "learning_rate": 0.002449195444415034,
+ "loss": 4.753599166870117,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.011197247542440891,
+ "learning_rate": 0.0024481589338805113,
+ "loss": 4.751479148864746,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.013252820819616318,
+ "learning_rate": 0.0024471216715166904,
+ "loss": 4.743409156799316,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.012366211041808128,
+ "learning_rate": 0.0024460836581592744,
+ "loss": 4.731625556945801,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.013906655833125114,
+ "learning_rate": 0.0024450448946445737,
+ "loss": 4.768153190612793,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.012118754908442497,
+ "learning_rate": 0.002444005381809499,
+ "loss": 4.74198579788208,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.01197550818324089,
+ "learning_rate": 0.002442965120491568,
+ "loss": 4.737289905548096,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.010640372522175312,
+ "learning_rate": 0.002441924111528901,
+ "loss": 4.7472686767578125,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.013553396798670292,
+ "learning_rate": 0.0024408823557602196,
+ "loss": 4.711215972900391,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.01393711194396019,
+ "learning_rate": 0.002439839854024847,
+ "loss": 4.720060348510742,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01589774526655674,
+ "learning_rate": 0.002438796607162709,
+ "loss": 4.695697784423828,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.0097908740863204,
+ "learning_rate": 0.0024377526160143303,
+ "loss": 4.729333877563477,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.011745489202439785,
+ "learning_rate": 0.0024367078814208363,
+ "loss": 4.722957134246826,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.009649662300944328,
+ "learning_rate": 0.0024356624042239498,
+ "loss": 4.727540493011475,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.008429240435361862,
+ "learning_rate": 0.002434616185265993,
+ "loss": 4.705052375793457,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.00984937883913517,
+ "learning_rate": 0.002433569225389887,
+ "loss": 4.721161365509033,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.009237290360033512,
+ "learning_rate": 0.0024325215254391466,
+ "loss": 4.724277496337891,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.008900545537471771,
+ "learning_rate": 0.002431473086257886,
+ "loss": 4.711522102355957,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.010003026574850082,
+ "learning_rate": 0.002430423908690813,
+ "loss": 4.728816032409668,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.011658324860036373,
+ "learning_rate": 0.0024293739935832316,
+ "loss": 4.716179847717285,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.01312185451388359,
+ "learning_rate": 0.0024283233417810393,
+ "loss": 4.7172346115112305,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.014306646771728992,
+ "learning_rate": 0.0024272719541307266,
+ "loss": 4.704054832458496,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.009844634681940079,
+ "learning_rate": 0.0024262198314793783,
+ "loss": 4.72379207611084,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.00903896801173687,
+ "learning_rate": 0.0024251669746746703,
+ "loss": 4.711025238037109,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.010476433672010899,
+ "learning_rate": 0.0024241133845648703,
+ "loss": 4.711664199829102,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.009128391742706299,
+ "learning_rate": 0.0024230590619988375,
+ "loss": 4.72395133972168,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.01228832732886076,
+ "learning_rate": 0.0024220040078260195,
+ "loss": 4.747347831726074,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.011916602961719036,
+ "learning_rate": 0.002420948222896455,
+ "loss": 4.696471691131592,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.012877700850367546,
+ "learning_rate": 0.0024198917080607706,
+ "loss": 4.731119632720947,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.012967764399945736,
+ "learning_rate": 0.0024188344641701814,
+ "loss": 4.688953876495361,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.011993546038866043,
+ "learning_rate": 0.002417776492076489,
+ "loss": 4.695237159729004,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.011188571341335773,
+ "learning_rate": 0.0024167177926320833,
+ "loss": 4.699850082397461,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.010556348599493504,
+ "learning_rate": 0.0024156583666899395,
+ "loss": 4.703063011169434,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.012023619376122952,
+ "learning_rate": 0.0024145982151036176,
+ "loss": 4.6918792724609375,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01266001258045435,
+ "learning_rate": 0.0024135373387272623,
+ "loss": 4.6653289794921875,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.012152725830674171,
+ "learning_rate": 0.0024124757384156024,
+ "loss": 4.682897567749023,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.014056944288313389,
+ "learning_rate": 0.002411413415023951,
+ "loss": 4.696136474609375,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.016826556995511055,
+ "learning_rate": 0.002410350369408202,
+ "loss": 4.701657295227051,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.015048800967633724,
+ "learning_rate": 0.0024092866024248323,
+ "loss": 4.691389083862305,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.012512968853116035,
+ "learning_rate": 0.0024082221149308997,
+ "loss": 4.677886009216309,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01596425101161003,
+ "learning_rate": 0.0024071569077840425,
+ "loss": 4.718337535858154,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01640196517109871,
+ "learning_rate": 0.002406090981842479,
+ "loss": 4.697303295135498,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.012347170151770115,
+ "learning_rate": 0.002405024337965007,
+ "loss": 4.696938991546631,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01213028747588396,
+ "learning_rate": 0.0024039569770110012,
+ "loss": 4.737306594848633,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01314542256295681,
+ "learning_rate": 0.002402888899840416,
+ "loss": 4.69948673248291,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.013764189556241035,
+ "learning_rate": 0.0024018201073137805,
+ "loss": 4.722800254821777,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.011147351004183292,
+ "learning_rate": 0.002400750600292203,
+ "loss": 4.694892883300781,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.015437614172697067,
+ "learning_rate": 0.0023996803796373653,
+ "loss": 4.704752445220947,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.016402097418904305,
+ "learning_rate": 0.0023986094462115255,
+ "loss": 4.67849063873291,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.011600594967603683,
+ "learning_rate": 0.002397537800877514,
+ "loss": 4.689641952514648,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.011864298023283482,
+ "learning_rate": 0.0023964654444987373,
+ "loss": 4.670377731323242,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01409872155636549,
+ "learning_rate": 0.002395392377939173,
+ "loss": 4.696022033691406,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.012106554582715034,
+ "learning_rate": 0.0023943186020633717,
+ "loss": 4.705615043640137,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.012310667894780636,
+ "learning_rate": 0.002393244117736455,
+ "loss": 4.702986717224121,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.013501073233783245,
+ "learning_rate": 0.002392168925824116,
+ "loss": 4.6700215339660645,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.013565126806497574,
+ "learning_rate": 0.0023910930271926166,
+ "loss": 4.699914455413818,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.013421077281236649,
+ "learning_rate": 0.00239001642270879,
+ "loss": 4.713191986083984,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.014408102259039879,
+ "learning_rate": 0.002388939113240035,
+ "loss": 4.655015468597412,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.013502835296094418,
+ "learning_rate": 0.0023878610996543224,
+ "loss": 4.6142258644104,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.014169570058584213,
+ "learning_rate": 0.002386782382820188,
+ "loss": 4.660243034362793,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.885881423950195,
+ "eval_runtime": 52.0315,
+ "eval_samples_per_second": 46.933,
+ "eval_steps_per_second": 1.48,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.014618083834648132,
+ "learning_rate": 0.002385702963606734,
+ "loss": 4.640046119689941,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01745622046291828,
+ "learning_rate": 0.0023846228428836278,
+ "loss": 4.641966819763184,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.014995258301496506,
+ "learning_rate": 0.002383542021521106,
+ "loss": 4.60922908782959,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.013278020545840263,
+ "learning_rate": 0.002382460500389964,
+ "loss": 4.652071952819824,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.014685257337987423,
+ "learning_rate": 0.002381378280361566,
+ "loss": 4.672174453735352,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.012737024575471878,
+ "learning_rate": 0.002380295362307836,
+ "loss": 4.612064361572266,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.012992842122912407,
+ "learning_rate": 0.002379211747101261,
+ "loss": 4.611995697021484,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.010210646316409111,
+ "learning_rate": 0.0023781274356148912,
+ "loss": 4.6055707931518555,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.010265499353408813,
+ "learning_rate": 0.0023770424287223365,
+ "loss": 4.637089729309082,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.010174009948968887,
+ "learning_rate": 0.002375956727297768,
+ "loss": 4.620571136474609,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.009512446820735931,
+ "learning_rate": 0.0023748703322159143,
+ "loss": 4.644092559814453,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.009849252179265022,
+ "learning_rate": 0.0023737832443520655,
+ "loss": 4.625055313110352,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.010493188165128231,
+ "learning_rate": 0.0023726954645820674,
+ "loss": 4.644410610198975,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.022181199863553047,
+ "learning_rate": 0.002371606993782326,
+ "loss": 4.593084335327148,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.011906919069588184,
+ "learning_rate": 0.0023705178328298016,
+ "loss": 4.6452836990356445,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.011874495074152946,
+ "learning_rate": 0.0023694279826020122,
+ "loss": 4.599322319030762,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.00989708211272955,
+ "learning_rate": 0.0023683374439770295,
+ "loss": 4.607342720031738,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.009256849065423012,
+ "learning_rate": 0.0023672462178334817,
+ "loss": 4.621583461761475,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.013187132775783539,
+ "learning_rate": 0.0023661543050505496,
+ "loss": 4.614629745483398,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.013505950570106506,
+ "learning_rate": 0.002365061706507968,
+ "loss": 4.629693984985352,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.012032733298838139,
+ "learning_rate": 0.0023639684230860225,
+ "loss": 4.671209335327148,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.01278337650001049,
+ "learning_rate": 0.002362874455665553,
+ "loss": 4.648874759674072,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.01123690977692604,
+ "learning_rate": 0.0023617798051279496,
+ "loss": 4.611470699310303,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.012564311735332012,
+ "learning_rate": 0.002360684472355151,
+ "loss": 4.659026145935059,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.01206638477742672,
+ "learning_rate": 0.002359588458229649,
+ "loss": 4.642566204071045,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.010300020687282085,
+ "learning_rate": 0.0023584917636344806,
+ "loss": 4.629585266113281,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.0102265365421772,
+ "learning_rate": 0.0023573943894532336,
+ "loss": 4.609250068664551,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.012440572492778301,
+ "learning_rate": 0.002356296336570042,
+ "loss": 4.636714458465576,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.011830241419374943,
+ "learning_rate": 0.0023551976058695887,
+ "loss": 4.59100866317749,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.011482699774205685,
+ "learning_rate": 0.0023540981982370997,
+ "loss": 4.620088577270508,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.012950294651091099,
+ "learning_rate": 0.0023529981145583483,
+ "loss": 4.578657627105713,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.012197710573673248,
+ "learning_rate": 0.002351897355719652,
+ "loss": 4.620813846588135,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.014839230105280876,
+ "learning_rate": 0.002350795922607872,
+ "loss": 4.615731239318848,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016670791432261467,
+ "learning_rate": 0.002349693816110414,
+ "loss": 4.655855178833008,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.014426100067794323,
+ "learning_rate": 0.0023485910371152247,
+ "loss": 4.660849571228027,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.014040560461580753,
+ "learning_rate": 0.0023474875865107936,
+ "loss": 4.634917736053467,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018269391730427742,
+ "learning_rate": 0.0023463834651861496,
+ "loss": 4.611809253692627,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.013855630531907082,
+ "learning_rate": 0.0023452786740308647,
+ "loss": 4.615225791931152,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.011505477130413055,
+ "learning_rate": 0.0023441732139350487,
+ "loss": 4.62592887878418,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.010166114196181297,
+ "learning_rate": 0.002343067085789351,
+ "loss": 4.6219024658203125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.009253021329641342,
+ "learning_rate": 0.0023419602904849596,
+ "loss": 4.658320426940918,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01027526892721653,
+ "learning_rate": 0.002340852828913598,
+ "loss": 4.589777946472168,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.008855022490024567,
+ "learning_rate": 0.002339744701967529,
+ "loss": 4.588204383850098,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.00965463649481535,
+ "learning_rate": 0.002338635910539551,
+ "loss": 4.648177146911621,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.009073720313608646,
+ "learning_rate": 0.0023375264555229956,
+ "loss": 4.585582733154297,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.010105383582413197,
+ "learning_rate": 0.0023364163378117317,
+ "loss": 4.612174987792969,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.015307556837797165,
+ "learning_rate": 0.002335305558300161,
+ "loss": 4.591984748840332,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016716254875063896,
+ "learning_rate": 0.002334194117883219,
+ "loss": 4.659285545349121,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.013281598687171936,
+ "learning_rate": 0.0023330820174563725,
+ "loss": 4.645829200744629,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.014416513964533806,
+ "learning_rate": 0.0023319692579156204,
+ "loss": 4.584893226623535,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.016326071694493294,
+ "learning_rate": 0.002330855840157494,
+ "loss": 4.633223533630371,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.012851824052631855,
+ "learning_rate": 0.002329741765079053,
+ "loss": 4.580986499786377,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.015444506891071796,
+ "learning_rate": 0.0023286270335778876,
+ "loss": 4.615891456604004,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.01391979493200779,
+ "learning_rate": 0.0023275116465521174,
+ "loss": 4.628403186798096,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.012138686142861843,
+ "learning_rate": 0.0023263956049003895,
+ "loss": 4.6259942054748535,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.013741672039031982,
+ "learning_rate": 0.0023252789095218785,
+ "loss": 4.6056742668151855,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01103319600224495,
+ "learning_rate": 0.002324161561316285,
+ "loss": 4.621670246124268,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.010333316400647163,
+ "learning_rate": 0.0023230435611838367,
+ "loss": 4.620484352111816,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.011063606478273869,
+ "learning_rate": 0.0023219249100252864,
+ "loss": 4.64952278137207,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.009984857402741909,
+ "learning_rate": 0.0023208056087419106,
+ "loss": 4.572571754455566,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.009986856952309608,
+ "learning_rate": 0.002319685658235511,
+ "loss": 4.570366382598877,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.010905452072620392,
+ "learning_rate": 0.0023185650594084103,
+ "loss": 4.603614807128906,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.012800061143934727,
+ "learning_rate": 0.002317443813163456,
+ "loss": 4.562705039978027,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.01252608560025692,
+ "learning_rate": 0.002316321920404015,
+ "loss": 4.599016189575195,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.011305592954158783,
+ "learning_rate": 0.002315199382033976,
+ "loss": 4.644617557525635,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.010594413615763187,
+ "learning_rate": 0.002314076198957749,
+ "loss": 4.569808006286621,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.009859138168394566,
+ "learning_rate": 0.002312952372080261,
+ "loss": 4.5890398025512695,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.010982917621731758,
+ "learning_rate": 0.00231182790230696,
+ "loss": 4.606348991394043,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.010252755135297775,
+ "learning_rate": 0.00231070279054381,
+ "loss": 4.58942174911499,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01110257301479578,
+ "learning_rate": 0.0023095770376972947,
+ "loss": 4.599823474884033,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01003815233707428,
+ "learning_rate": 0.0023084506446744103,
+ "loss": 4.569343090057373,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.010517258197069168,
+ "learning_rate": 0.0023073236123826745,
+ "loss": 4.646301746368408,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.804611682891846,
+ "eval_runtime": 51.6279,
+ "eval_samples_per_second": 47.3,
+ "eval_steps_per_second": 1.491,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.011439066380262375,
+ "learning_rate": 0.0023061959417301145,
+ "loss": 4.563335418701172,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.01073157787322998,
+ "learning_rate": 0.002305067633625275,
+ "loss": 4.540966987609863,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.011391228064894676,
+ "learning_rate": 0.0023039386889772136,
+ "loss": 4.555583953857422,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.012814595364034176,
+ "learning_rate": 0.0023028091086955,
+ "loss": 4.517425537109375,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.012994988821446896,
+ "learning_rate": 0.002301678893690218,
+ "loss": 4.538477897644043,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.01460991520434618,
+ "learning_rate": 0.00230054804487196,
+ "loss": 4.571145057678223,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.013036075979471207,
+ "learning_rate": 0.0022994165631518306,
+ "loss": 4.519233703613281,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.011917057447135448,
+ "learning_rate": 0.0022982844494414444,
+ "loss": 4.5071587562561035,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.013928866013884544,
+ "learning_rate": 0.0022971517046529257,
+ "loss": 4.531422138214111,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01407038513571024,
+ "learning_rate": 0.0022960183296989053,
+ "loss": 4.511205196380615,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.019396567717194557,
+ "learning_rate": 0.0022948843254925245,
+ "loss": 4.530007362365723,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.014425835572183132,
+ "learning_rate": 0.002293749692947428,
+ "loss": 4.530138969421387,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.01600247249007225,
+ "learning_rate": 0.00229261443297777,
+ "loss": 4.5643391609191895,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.018238037824630737,
+ "learning_rate": 0.0022914785464982095,
+ "loss": 4.540779113769531,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.015427535399794579,
+ "learning_rate": 0.0022903420344239085,
+ "loss": 4.506287097930908,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.016816625371575356,
+ "learning_rate": 0.0022892048976705347,
+ "loss": 4.537919521331787,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.01999717950820923,
+ "learning_rate": 0.002288067137154259,
+ "loss": 4.558426380157471,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.020884918048977852,
+ "learning_rate": 0.002286928753791753,
+ "loss": 4.592915058135986,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.01620141603052616,
+ "learning_rate": 0.002285789748500194,
+ "loss": 4.553356170654297,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.018495604395866394,
+ "learning_rate": 0.0022846501221972567,
+ "loss": 4.573139190673828,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.0156320258975029,
+ "learning_rate": 0.0022835098758011174,
+ "loss": 4.585921764373779,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.014515317976474762,
+ "learning_rate": 0.002282369010230452,
+ "loss": 4.5532002449035645,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.014826556667685509,
+ "learning_rate": 0.0022812275264044355,
+ "loss": 4.560998916625977,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.013240831904113293,
+ "learning_rate": 0.00228008542524274,
+ "loss": 4.531787872314453,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.013814719393849373,
+ "learning_rate": 0.0022789427076655372,
+ "loss": 4.611872673034668,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.012246561236679554,
+ "learning_rate": 0.0022777993745934934,
+ "loss": 4.5565385818481445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01206218171864748,
+ "learning_rate": 0.002276655426947771,
+ "loss": 4.570852279663086,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01007329672574997,
+ "learning_rate": 0.0022755108656500275,
+ "loss": 4.553400039672852,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.010361720807850361,
+ "learning_rate": 0.0022743656916224173,
+ "loss": 4.527509689331055,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.009456256404519081,
+ "learning_rate": 0.0022732199057875846,
+ "loss": 4.520236015319824,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.009310049936175346,
+ "learning_rate": 0.002272073509068669,
+ "loss": 4.518343925476074,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.00927449855953455,
+ "learning_rate": 0.0022709265023893015,
+ "loss": 4.559690475463867,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.008658594451844692,
+ "learning_rate": 0.002269778886673605,
+ "loss": 4.513876914978027,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.008928220719099045,
+ "learning_rate": 0.0022686306628461918,
+ "loss": 4.545050621032715,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.008046358823776245,
+ "learning_rate": 0.002267481831832166,
+ "loss": 4.54866361618042,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.008318924345076084,
+ "learning_rate": 0.00226633239455712,
+ "loss": 4.546968936920166,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.008808869868516922,
+ "learning_rate": 0.0022651823519471343,
+ "loss": 4.50877571105957,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.008767383173108101,
+ "learning_rate": 0.0022640317049287774,
+ "loss": 4.57666540145874,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.009351354092359543,
+ "learning_rate": 0.0022628804544291047,
+ "loss": 4.511702537536621,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.010137057863175869,
+ "learning_rate": 0.002261728601375659,
+ "loss": 4.539676189422607,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.00882046204060316,
+ "learning_rate": 0.002260576146696466,
+ "loss": 4.533196449279785,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.009354782290756702,
+ "learning_rate": 0.0022594230913200388,
+ "loss": 4.515488624572754,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.009241201914846897,
+ "learning_rate": 0.002258269436175372,
+ "loss": 4.542706489562988,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.009252829477190971,
+ "learning_rate": 0.002257115182191947,
+ "loss": 4.592660903930664,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.011243131943047047,
+ "learning_rate": 0.0022559603302997225,
+ "loss": 4.527804851531982,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.011032823473215103,
+ "learning_rate": 0.0022548048814291437,
+ "loss": 4.540122032165527,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.010495275259017944,
+ "learning_rate": 0.0022536488365111347,
+ "loss": 4.532835006713867,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.009992090053856373,
+ "learning_rate": 0.0022524921964770994,
+ "loss": 4.521217346191406,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01141198631376028,
+ "learning_rate": 0.0022513349622589225,
+ "loss": 4.519709587097168,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.013087889179587364,
+ "learning_rate": 0.002250177134788966,
+ "loss": 4.547289848327637,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.01328477542847395,
+ "learning_rate": 0.0022490187150000706,
+ "loss": 4.527865409851074,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01033794041723013,
+ "learning_rate": 0.002247859703825555,
+ "loss": 4.54740047454834,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.010522300377488136,
+ "learning_rate": 0.002246700102199213,
+ "loss": 4.538840293884277,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.012151090428233147,
+ "learning_rate": 0.0022455399110553145,
+ "loss": 4.536865234375,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.010553102008998394,
+ "learning_rate": 0.0022443791313286054,
+ "loss": 4.5348334312438965,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.011433709412813187,
+ "learning_rate": 0.0022432177639543035,
+ "loss": 4.510526657104492,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01189691387116909,
+ "learning_rate": 0.002242055809868103,
+ "loss": 4.530807018280029,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.010031254962086678,
+ "learning_rate": 0.0022408932700061683,
+ "loss": 4.496850967407227,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.010129331611096859,
+ "learning_rate": 0.002239730145305138,
+ "loss": 4.551287651062012,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.010430586524307728,
+ "learning_rate": 0.0022385664367021185,
+ "loss": 4.542672157287598,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.011774476617574692,
+ "learning_rate": 0.002237402145134691,
+ "loss": 4.537034034729004,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.01184659544378519,
+ "learning_rate": 0.002236237271540903,
+ "loss": 4.521193504333496,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.011327247135341167,
+ "learning_rate": 0.0022350718168592724,
+ "loss": 4.49426794052124,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.010767918080091476,
+ "learning_rate": 0.0022339057820287855,
+ "loss": 4.531225204467773,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.011195424944162369,
+ "learning_rate": 0.002232739167988894,
+ "loss": 4.492671012878418,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.010394908487796783,
+ "learning_rate": 0.0022315719756795197,
+ "loss": 4.523616790771484,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.010277998633682728,
+ "learning_rate": 0.002230404206041047,
+ "loss": 4.487832069396973,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.010955783538520336,
+ "learning_rate": 0.0022292358600143275,
+ "loss": 4.526865005493164,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.010611076839268208,
+ "learning_rate": 0.002228066938540676,
+ "loss": 4.498682022094727,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.010008828714489937,
+ "learning_rate": 0.0022268974425618724,
+ "loss": 4.524019241333008,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.008973540738224983,
+ "learning_rate": 0.002225727373020157,
+ "loss": 4.512996673583984,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.011754963546991348,
+ "learning_rate": 0.002224556730858235,
+ "loss": 4.496376991271973,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.744190216064453,
+ "eval_runtime": 51.6572,
+ "eval_samples_per_second": 47.273,
+ "eval_steps_per_second": 1.491,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.012094628997147083,
+ "learning_rate": 0.0022233855170192703,
+ "loss": 4.443447113037109,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.014347990974783897,
+ "learning_rate": 0.00222221373244689,
+ "loss": 4.485690116882324,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.021286984905600548,
+ "learning_rate": 0.0022210413780851786,
+ "loss": 4.487505912780762,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.020273910835385323,
+ "learning_rate": 0.0022198684548786827,
+ "loss": 4.473119735717773,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.017642971128225327,
+ "learning_rate": 0.002218694963772403,
+ "loss": 4.440615653991699,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.02075147069990635,
+ "learning_rate": 0.0022175209057118017,
+ "loss": 4.527026176452637,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016479508951306343,
+ "learning_rate": 0.002216346281642796,
+ "loss": 4.508321762084961,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.017978593707084656,
+ "learning_rate": 0.002215171092511759,
+ "loss": 4.500985145568848,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.01739897020161152,
+ "learning_rate": 0.002213995339265519,
+ "loss": 4.474225044250488,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.015556776896119118,
+ "learning_rate": 0.0022128190228513597,
+ "loss": 4.444849014282227,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.016166578978300095,
+ "learning_rate": 0.002211642144217018,
+ "loss": 4.5026726722717285,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.01534261740744114,
+ "learning_rate": 0.0022104647043106832,
+ "loss": 4.484527587890625,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.013517106883227825,
+ "learning_rate": 0.002209286704080998,
+ "loss": 4.453907012939453,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.012562787160277367,
+ "learning_rate": 0.0022081081444770556,
+ "loss": 4.471528053283691,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.009947682730853558,
+ "learning_rate": 0.0022069290264483997,
+ "loss": 4.504153728485107,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01169571466743946,
+ "learning_rate": 0.002205749350945026,
+ "loss": 4.447345733642578,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.011394352652132511,
+ "learning_rate": 0.0022045691189173755,
+ "loss": 4.447491645812988,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.011168593540787697,
+ "learning_rate": 0.0022033883313163417,
+ "loss": 4.441639423370361,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.009503341279923916,
+ "learning_rate": 0.002202206989093263,
+ "loss": 4.461601257324219,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.0112199941650033,
+ "learning_rate": 0.002201025093199926,
+ "loss": 4.444095134735107,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.010505333542823792,
+ "learning_rate": 0.002199842644588562,
+ "loss": 4.460987091064453,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.009494813159108162,
+ "learning_rate": 0.0021986596442118496,
+ "loss": 4.460158348083496,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.01063452661037445,
+ "learning_rate": 0.00219747609302291,
+ "loss": 4.489625453948975,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.012194888666272163,
+ "learning_rate": 0.0021962919919753087,
+ "loss": 4.441605567932129,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.010613647289574146,
+ "learning_rate": 0.0021951073420230557,
+ "loss": 4.475821018218994,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.009643700905144215,
+ "learning_rate": 0.002193922144120601,
+ "loss": 4.467228889465332,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.00912613607943058,
+ "learning_rate": 0.0021927363992228373,
+ "loss": 4.455435752868652,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.009371735155582428,
+ "learning_rate": 0.0021915501082850988,
+ "loss": 4.479915142059326,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.008617875166237354,
+ "learning_rate": 0.002190363272263158,
+ "loss": 4.415642738342285,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.009571383707225323,
+ "learning_rate": 0.002189175892113227,
+ "loss": 4.439854621887207,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.00782635249197483,
+ "learning_rate": 0.002187987968791957,
+ "loss": 4.484311103820801,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.008408057503402233,
+ "learning_rate": 0.0021867995032564358,
+ "loss": 4.46577262878418,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.009641910903155804,
+ "learning_rate": 0.00218561049646419,
+ "loss": 4.437408447265625,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.008498182520270348,
+ "learning_rate": 0.0021844209493731784,
+ "loss": 4.442599296569824,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.009289798326790333,
+ "learning_rate": 0.0021832308629418,
+ "loss": 4.517641067504883,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.009360183961689472,
+ "learning_rate": 0.0021820402381288848,
+ "loss": 4.474143981933594,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.008764738216996193,
+ "learning_rate": 0.0021808490758936983,
+ "loss": 4.4376702308654785,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.009673701599240303,
+ "learning_rate": 0.0021796573771959375,
+ "loss": 4.435220718383789,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.010852573439478874,
+ "learning_rate": 0.0021784651429957334,
+ "loss": 4.465187072753906,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.011292314156889915,
+ "learning_rate": 0.0021772723742536464,
+ "loss": 4.458176136016846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01071334257721901,
+ "learning_rate": 0.0021760790719306702,
+ "loss": 4.425498962402344,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.01505239401012659,
+ "learning_rate": 0.0021748852369882265,
+ "loss": 4.463129043579102,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.011774689890444279,
+ "learning_rate": 0.0021736908703881663,
+ "loss": 4.462577819824219,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.011241118423640728,
+ "learning_rate": 0.002172495973092769,
+ "loss": 4.452583312988281,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.01072910986840725,
+ "learning_rate": 0.002171300546064742,
+ "loss": 4.466248035430908,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01021308358758688,
+ "learning_rate": 0.002170104590267219,
+ "loss": 4.485857963562012,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.012630444951355457,
+ "learning_rate": 0.00216890810666376,
+ "loss": 4.442112922668457,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.013602569699287415,
+ "learning_rate": 0.0021677110962183515,
+ "loss": 4.45171594619751,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.013808641582727432,
+ "learning_rate": 0.0021665135598954014,
+ "loss": 4.468818187713623,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.011510699987411499,
+ "learning_rate": 0.002165315498659744,
+ "loss": 4.439584255218506,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.010502424091100693,
+ "learning_rate": 0.0021641169134766353,
+ "loss": 4.478377342224121,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.009953407570719719,
+ "learning_rate": 0.0021629178053117542,
+ "loss": 4.474404335021973,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.011158930137753487,
+ "learning_rate": 0.0021617181751312,
+ "loss": 4.4549560546875,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.010431940667331219,
+ "learning_rate": 0.002160518023901493,
+ "loss": 4.444911003112793,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.009100335650146008,
+ "learning_rate": 0.0021593173525895734,
+ "loss": 4.428071975708008,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.009362847544252872,
+ "learning_rate": 0.0021581161621628013,
+ "loss": 4.462827682495117,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.009893129579722881,
+ "learning_rate": 0.0021569144535889535,
+ "loss": 4.426912307739258,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.010524468496441841,
+ "learning_rate": 0.0021557122278362244,
+ "loss": 4.435606002807617,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.010621445253491402,
+ "learning_rate": 0.002154509485873226,
+ "loss": 4.454901695251465,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.010836896486580372,
+ "learning_rate": 0.0021533062286689855,
+ "loss": 4.4441728591918945,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.011403030715882778,
+ "learning_rate": 0.0021521024571929457,
+ "loss": 4.483884811401367,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.012522426433861256,
+ "learning_rate": 0.0021508981724149643,
+ "loss": 4.464006423950195,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.014799179509282112,
+ "learning_rate": 0.0021496933753053105,
+ "loss": 4.460402011871338,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.012453878298401833,
+ "learning_rate": 0.0021484880668346685,
+ "loss": 4.456899642944336,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.012270379811525345,
+ "learning_rate": 0.002147282247974133,
+ "loss": 4.477325439453125,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01094560045748949,
+ "learning_rate": 0.0021460759196952105,
+ "loss": 4.422362327575684,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.009995176456868649,
+ "learning_rate": 0.0021448690829698183,
+ "loss": 4.405127048492432,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.014935536310076714,
+ "learning_rate": 0.002143661738770283,
+ "loss": 4.445840358734131,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.011492030695080757,
+ "learning_rate": 0.0021424538880693393,
+ "loss": 4.459555625915527,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.008793608285486698,
+ "learning_rate": 0.002141245531840132,
+ "loss": 4.436631202697754,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.008894025348126888,
+ "learning_rate": 0.0021400366710562103,
+ "loss": 4.473924160003662,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.010324863716959953,
+ "learning_rate": 0.0021388273066915327,
+ "loss": 4.475680351257324,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.688324451446533,
+ "eval_runtime": 54.4515,
+ "eval_samples_per_second": 44.847,
+ "eval_steps_per_second": 1.414,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.009235628880560398,
+ "learning_rate": 0.0021376174397204615,
+ "loss": 4.370064735412598,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.008818200789391994,
+ "learning_rate": 0.0021364070711177643,
+ "loss": 4.371678352355957,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.010333757847547531,
+ "learning_rate": 0.0021351962018586147,
+ "loss": 4.3896379470825195,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.011938809417188168,
+ "learning_rate": 0.002133984832918587,
+ "loss": 4.361343860626221,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.010634614154696465,
+ "learning_rate": 0.00213277296527366,
+ "loss": 4.362449645996094,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.008545687422156334,
+ "learning_rate": 0.0021315605999002126,
+ "loss": 4.3685503005981445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.009159795008599758,
+ "learning_rate": 0.002130347737775027,
+ "loss": 4.379969596862793,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.010464153252542019,
+ "learning_rate": 0.0021291343798752835,
+ "loss": 4.376113414764404,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.011010527610778809,
+ "learning_rate": 0.0021279205271785622,
+ "loss": 4.390711784362793,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.010350867174565792,
+ "learning_rate": 0.002126706180662843,
+ "loss": 4.399683952331543,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.009708237834274769,
+ "learning_rate": 0.002125491341306503,
+ "loss": 4.363435745239258,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.010087115690112114,
+ "learning_rate": 0.002124276010088316,
+ "loss": 4.366623878479004,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.009712349623441696,
+ "learning_rate": 0.002123060187987453,
+ "loss": 4.360814571380615,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.010174986906349659,
+ "learning_rate": 0.0021218438759834785,
+ "loss": 4.347475051879883,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.010195785202085972,
+ "learning_rate": 0.0021206270750563543,
+ "loss": 4.399251461029053,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.010786959901452065,
+ "learning_rate": 0.0021194097861864354,
+ "loss": 4.397933006286621,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.011476189829409122,
+ "learning_rate": 0.002118192010354468,
+ "loss": 4.3748369216918945,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.011571013368666172,
+ "learning_rate": 0.0021169737485415927,
+ "loss": 4.388349533081055,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.011140935122966766,
+ "learning_rate": 0.002115755001729341,
+ "loss": 4.398018836975098,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.013216734863817692,
+ "learning_rate": 0.0021145357708996365,
+ "loss": 4.406947135925293,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.013385126367211342,
+ "learning_rate": 0.0021133160570347893,
+ "loss": 4.41457462310791,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.011784231290221214,
+ "learning_rate": 0.002112095861117502,
+ "loss": 4.403835296630859,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.009813498705625534,
+ "learning_rate": 0.0021108751841308645,
+ "loss": 4.3865647315979,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.010297936387360096,
+ "learning_rate": 0.0021096540270583533,
+ "loss": 4.395583152770996,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.011965147219598293,
+ "learning_rate": 0.002108432390883834,
+ "loss": 4.385443687438965,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.012212193571031094,
+ "learning_rate": 0.0021072102765915556,
+ "loss": 4.422321319580078,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.009840975515544415,
+ "learning_rate": 0.0021059876851661544,
+ "loss": 4.3568902015686035,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.013039806857705116,
+ "learning_rate": 0.002104764617592649,
+ "loss": 4.314867973327637,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.015091902576386929,
+ "learning_rate": 0.0021035410748564444,
+ "loss": 4.392338752746582,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.015448705293238163,
+ "learning_rate": 0.0021023170579433257,
+ "loss": 4.379609107971191,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.018393104895949364,
+ "learning_rate": 0.0021010925678394622,
+ "loss": 4.418291091918945,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.01866372860968113,
+ "learning_rate": 0.002099867605531402,
+ "loss": 4.4006452560424805,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.01819106750190258,
+ "learning_rate": 0.002098642172006077,
+ "loss": 4.433975696563721,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.019774602726101875,
+ "learning_rate": 0.0020974162682507952,
+ "loss": 4.417821884155273,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.016885802149772644,
+ "learning_rate": 0.0020961898952532457,
+ "loss": 4.397073268890381,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.015750881284475327,
+ "learning_rate": 0.0020949630540014956,
+ "loss": 4.4356608390808105,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.015696514397859573,
+ "learning_rate": 0.002093735745483988,
+ "loss": 4.422449588775635,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.017195021733641624,
+ "learning_rate": 0.0020925079706895435,
+ "loss": 4.443209648132324,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.015864048153162003,
+ "learning_rate": 0.0020912797306073577,
+ "loss": 4.4340691566467285,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.009857800789177418,
+ "learning_rate": 0.0020900510262270022,
+ "loss": 4.37935209274292,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.013739476911723614,
+ "learning_rate": 0.0020888218585384216,
+ "loss": 4.412148475646973,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.011667690239846706,
+ "learning_rate": 0.0020875922285319344,
+ "loss": 4.391700744628906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.012323460541665554,
+ "learning_rate": 0.0020863621371982302,
+ "loss": 4.423880577087402,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.010353626683354378,
+ "learning_rate": 0.0020851315855283724,
+ "loss": 4.427382469177246,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.0101751908659935,
+ "learning_rate": 0.002083900574513794,
+ "loss": 4.415317535400391,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.01126549206674099,
+ "learning_rate": 0.002082669105146298,
+ "loss": 4.402498245239258,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.011096738278865814,
+ "learning_rate": 0.0020814371784180575,
+ "loss": 4.433567047119141,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.010944445617496967,
+ "learning_rate": 0.0020802047953216137,
+ "loss": 4.406990051269531,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.008965318091213703,
+ "learning_rate": 0.0020789719568498747,
+ "loss": 4.3721771240234375,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.00828113779425621,
+ "learning_rate": 0.002077738663996117,
+ "loss": 4.389828205108643,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.008599454537034035,
+ "learning_rate": 0.002076504917753982,
+ "loss": 4.411396026611328,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.007713889703154564,
+ "learning_rate": 0.0020752707191174762,
+ "loss": 4.355343341827393,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.00912355538457632,
+ "learning_rate": 0.0020740360690809723,
+ "loss": 4.423413276672363,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.008269775658845901,
+ "learning_rate": 0.0020728009686392045,
+ "loss": 4.390920639038086,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.007461601868271828,
+ "learning_rate": 0.002071565418787272,
+ "loss": 4.365501403808594,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.007951333187520504,
+ "learning_rate": 0.0020703294205206336,
+ "loss": 4.378884315490723,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.007816503755748272,
+ "learning_rate": 0.002069092974835112,
+ "loss": 4.4241743087768555,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.00833756010979414,
+ "learning_rate": 0.0020678560827268886,
+ "loss": 4.391102313995361,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.008241684176027775,
+ "learning_rate": 0.0020666187451925055,
+ "loss": 4.374970436096191,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.008245996199548244,
+ "learning_rate": 0.002065380963228862,
+ "loss": 4.380222320556641,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.007378706708550453,
+ "learning_rate": 0.0020641427378332183,
+ "loss": 4.391451358795166,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.007594958413392305,
+ "learning_rate": 0.002062904070003189,
+ "loss": 4.372774124145508,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.0076752156019210815,
+ "learning_rate": 0.002061664960736747,
+ "loss": 4.384668827056885,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.008727089501917362,
+ "learning_rate": 0.00206042541103222,
+ "loss": 4.363401412963867,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.007616396062076092,
+ "learning_rate": 0.002059185421888291,
+ "loss": 4.413188934326172,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.008922954089939594,
+ "learning_rate": 0.0020579449943039967,
+ "loss": 4.383417129516602,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.009063739329576492,
+ "learning_rate": 0.002056704129278727,
+ "loss": 4.385283470153809,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.008912574499845505,
+ "learning_rate": 0.002055462827812225,
+ "loss": 4.378478050231934,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.008317525498569012,
+ "learning_rate": 0.0020542210909045844,
+ "loss": 4.367859840393066,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.009347227402031422,
+ "learning_rate": 0.0020529789195562506,
+ "loss": 4.38238000869751,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.011030065827071667,
+ "learning_rate": 0.0020517363147680183,
+ "loss": 4.372624397277832,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.011519700288772583,
+ "learning_rate": 0.002050493277541032,
+ "loss": 4.3906121253967285,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.641542434692383,
+ "eval_runtime": 51.6342,
+ "eval_samples_per_second": 47.294,
+ "eval_steps_per_second": 1.491,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.009694516658782959,
+ "learning_rate": 0.0020492498088767836,
+ "loss": 4.32808256149292,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.0097952950745821,
+ "learning_rate": 0.002048005909777114,
+ "loss": 4.280096054077148,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.010393359698355198,
+ "learning_rate": 0.0020467615812442106,
+ "loss": 4.317824363708496,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.00977372657507658,
+ "learning_rate": 0.0020455168242806054,
+ "loss": 4.3445329666137695,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.008568719029426575,
+ "learning_rate": 0.0020442716398891775,
+ "loss": 4.318026542663574,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.009732510894536972,
+ "learning_rate": 0.0020430260290731487,
+ "loss": 4.330609321594238,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.010429969057440758,
+ "learning_rate": 0.0020417799928360864,
+ "loss": 4.316794395446777,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.009292383678257465,
+ "learning_rate": 0.002040533532181898,
+ "loss": 4.317420959472656,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.011222455650568008,
+ "learning_rate": 0.0020392866481148356,
+ "loss": 4.292610168457031,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.013591725379228592,
+ "learning_rate": 0.0020380393416394894,
+ "loss": 4.325689315795898,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.013572026044130325,
+ "learning_rate": 0.002036791613760794,
+ "loss": 4.334960460662842,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.012986521236598492,
+ "learning_rate": 0.0020355434654840197,
+ "loss": 4.33829402923584,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.013321424834430218,
+ "learning_rate": 0.0020342948978147777,
+ "loss": 4.2926926612854,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.012307215481996536,
+ "learning_rate": 0.0020330459117590153,
+ "loss": 4.32757568359375,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.012909360229969025,
+ "learning_rate": 0.0020317965083230197,
+ "loss": 4.362889289855957,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017746610566973686,
+ "learning_rate": 0.0020305466885134115,
+ "loss": 4.3426313400268555,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.01696564070880413,
+ "learning_rate": 0.002029296453337148,
+ "loss": 4.3298516273498535,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.013348255306482315,
+ "learning_rate": 0.002028045803801522,
+ "loss": 4.360220432281494,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.013225145637989044,
+ "learning_rate": 0.002026794740914158,
+ "loss": 4.346250057220459,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.01569020375609398,
+ "learning_rate": 0.002025543265683016,
+ "loss": 4.308025360107422,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.018612321466207504,
+ "learning_rate": 0.002024291379116385,
+ "loss": 4.330190181732178,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.01619943603873253,
+ "learning_rate": 0.00202303908222289,
+ "loss": 4.339552402496338,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.015385213308036327,
+ "learning_rate": 0.0020217863760114826,
+ "loss": 4.3412184715271,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.018517112359404564,
+ "learning_rate": 0.002020533261491446,
+ "loss": 4.3289313316345215,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.014868509024381638,
+ "learning_rate": 0.0020192797396723903,
+ "loss": 4.359846115112305,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.014240248128771782,
+ "learning_rate": 0.0020180258115642578,
+ "loss": 4.369707107543945,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.018113158643245697,
+ "learning_rate": 0.0020167714781773143,
+ "loss": 4.362668514251709,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.01802460104227066,
+ "learning_rate": 0.002015516740522154,
+ "loss": 4.344742298126221,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.01414549071341753,
+ "learning_rate": 0.0020142615996096964,
+ "loss": 4.358080863952637,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.012157700955867767,
+ "learning_rate": 0.002013006056451186,
+ "loss": 4.347763538360596,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.011268534697592258,
+ "learning_rate": 0.0020117501120581905,
+ "loss": 4.335859298706055,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.010395628400146961,
+ "learning_rate": 0.002010493767442602,
+ "loss": 4.341609001159668,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.012147950939834118,
+ "learning_rate": 0.002009237023616635,
+ "loss": 4.328502655029297,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.00945813488215208,
+ "learning_rate": 0.0020079798815928252,
+ "loss": 4.311186790466309,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.009489500895142555,
+ "learning_rate": 0.0020067223423840294,
+ "loss": 4.371470928192139,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.009360029362142086,
+ "learning_rate": 0.0020054644070034234,
+ "loss": 4.33941125869751,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.010032779537141323,
+ "learning_rate": 0.0020042060764645037,
+ "loss": 4.365797519683838,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.009367046877741814,
+ "learning_rate": 0.002002947351781084,
+ "loss": 4.29443359375,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.008369733579456806,
+ "learning_rate": 0.0020016882339672967,
+ "loss": 4.294233322143555,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.010394785553216934,
+ "learning_rate": 0.0020004287240375893,
+ "loss": 4.352355003356934,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01068816613405943,
+ "learning_rate": 0.001999168823006728,
+ "loss": 4.342516899108887,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.009975973516702652,
+ "learning_rate": 0.0019979085318897905,
+ "loss": 4.346763610839844,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.009618151932954788,
+ "learning_rate": 0.001996647851702171,
+ "loss": 4.329115867614746,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.008399694226682186,
+ "learning_rate": 0.001995386783459577,
+ "loss": 4.28736686706543,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.008235268294811249,
+ "learning_rate": 0.001994125328178029,
+ "loss": 4.339872360229492,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.00836547277867794,
+ "learning_rate": 0.0019928634868738577,
+ "loss": 4.328156471252441,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.009619626216590405,
+ "learning_rate": 0.001991601260563706,
+ "loss": 4.337394714355469,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.009814038872718811,
+ "learning_rate": 0.0019903386502645275,
+ "loss": 4.325926780700684,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.009588954970240593,
+ "learning_rate": 0.001989075656993584,
+ "loss": 4.286477088928223,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.010471943765878677,
+ "learning_rate": 0.001987812281768447,
+ "loss": 4.322466850280762,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.010239035822451115,
+ "learning_rate": 0.001986548525606994,
+ "loss": 4.3487653732299805,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.008935568854212761,
+ "learning_rate": 0.001985284389527411,
+ "loss": 4.303601264953613,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.008624759502708912,
+ "learning_rate": 0.00198401987454819,
+ "loss": 4.34036922454834,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.009296547621488571,
+ "learning_rate": 0.0019827549816881277,
+ "loss": 4.273357391357422,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.009071138687431812,
+ "learning_rate": 0.0019814897119663248,
+ "loss": 4.321051597595215,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.008341389708220959,
+ "learning_rate": 0.0019802240664021876,
+ "loss": 4.343242645263672,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.007881715893745422,
+ "learning_rate": 0.0019789580460154227,
+ "loss": 4.32067346572876,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.008750553242862225,
+ "learning_rate": 0.0019776916518260404,
+ "loss": 4.330263614654541,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.008591460064053535,
+ "learning_rate": 0.001976424884854351,
+ "loss": 4.311957359313965,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.008084543980658054,
+ "learning_rate": 0.001975157746120966,
+ "loss": 4.331319808959961,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.008036594837903976,
+ "learning_rate": 0.001973890236646797,
+ "loss": 4.382656097412109,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.00805772002786398,
+ "learning_rate": 0.0019726223574530527,
+ "loss": 4.297247886657715,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.009019420482218266,
+ "learning_rate": 0.0019713541095612405,
+ "loss": 4.326526165008545,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.009953883476555347,
+ "learning_rate": 0.0019700854939931653,
+ "loss": 4.327356338500977,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.009340492077171803,
+ "learning_rate": 0.0019688165117709273,
+ "loss": 4.343993186950684,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.010395721532404423,
+ "learning_rate": 0.0019675471639169226,
+ "loss": 4.284934997558594,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.009946230798959732,
+ "learning_rate": 0.0019662774514538415,
+ "loss": 4.288936614990234,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.010981100611388683,
+ "learning_rate": 0.0019650073754046695,
+ "loss": 4.339935779571533,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.009688783437013626,
+ "learning_rate": 0.0019637369367926827,
+ "loss": 4.381425857543945,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.009442255832254887,
+ "learning_rate": 0.001962466136641452,
+ "loss": 4.304185390472412,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.009371291846036911,
+ "learning_rate": 0.0019611949759748364,
+ "loss": 4.306120872497559,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.009544814005494118,
+ "learning_rate": 0.001959923455816988,
+ "loss": 4.326251983642578,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.602745532989502,
+ "eval_runtime": 53.5041,
+ "eval_samples_per_second": 45.641,
+ "eval_steps_per_second": 1.439,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.01010364294052124,
+ "learning_rate": 0.0019586515771923474,
+ "loss": 4.24491024017334,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.012722729705274105,
+ "learning_rate": 0.0019573793411256446,
+ "loss": 4.257739543914795,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.01307204831391573,
+ "learning_rate": 0.0019561067486418967,
+ "loss": 4.291684150695801,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.011215966194868088,
+ "learning_rate": 0.0019548338007664098,
+ "loss": 4.237947463989258,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.012943019159138203,
+ "learning_rate": 0.0019535604985247727,
+ "loss": 4.242859363555908,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.01337282732129097,
+ "learning_rate": 0.0019522868429428636,
+ "loss": 4.238934516906738,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.013697991147637367,
+ "learning_rate": 0.0019510128350468438,
+ "loss": 4.285760402679443,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.012182159349322319,
+ "learning_rate": 0.001949738475863158,
+ "loss": 4.290044784545898,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.012620339170098305,
+ "learning_rate": 0.0019484637664185346,
+ "loss": 4.267187118530273,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.011284958571195602,
+ "learning_rate": 0.0019471887077399837,
+ "loss": 4.26662540435791,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.01161450706422329,
+ "learning_rate": 0.001945913300854797,
+ "loss": 4.283192157745361,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.010022968985140324,
+ "learning_rate": 0.001944637546790548,
+ "loss": 4.295843124389648,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.010341878980398178,
+ "learning_rate": 0.001943361446575087,
+ "loss": 4.295636177062988,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.011549185030162334,
+ "learning_rate": 0.0019420850012365457,
+ "loss": 4.288768768310547,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.012518810108304024,
+ "learning_rate": 0.0019408082118033331,
+ "loss": 4.229338645935059,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.012035328894853592,
+ "learning_rate": 0.0019395310793041353,
+ "loss": 4.258639335632324,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.011193782091140747,
+ "learning_rate": 0.0019382536047679146,
+ "loss": 4.239264488220215,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.009930136613547802,
+ "learning_rate": 0.0019369757892239096,
+ "loss": 4.252286434173584,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.00953229982405901,
+ "learning_rate": 0.0019356976337016327,
+ "loss": 4.293435096740723,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.010618340224027634,
+ "learning_rate": 0.0019344191392308716,
+ "loss": 4.235588550567627,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.008806651458144188,
+ "learning_rate": 0.0019331403068416856,
+ "loss": 4.26716423034668,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.008934796787798405,
+ "learning_rate": 0.0019318611375644065,
+ "loss": 4.258462905883789,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.009283283725380898,
+ "learning_rate": 0.0019305816324296386,
+ "loss": 4.257868766784668,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.009224812500178814,
+ "learning_rate": 0.0019293017924682555,
+ "loss": 4.258486747741699,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.008957969956099987,
+ "learning_rate": 0.0019280216187114016,
+ "loss": 4.309563636779785,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.010284051299095154,
+ "learning_rate": 0.0019267411121904894,
+ "loss": 4.275271415710449,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.009167762473225594,
+ "learning_rate": 0.0019254602739372001,
+ "loss": 4.238452911376953,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.009411636739969254,
+ "learning_rate": 0.0019241791049834816,
+ "loss": 4.2735595703125,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.009044171310961246,
+ "learning_rate": 0.0019228976063615497,
+ "loss": 4.27017879486084,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.008840068243443966,
+ "learning_rate": 0.0019216157791038834,
+ "loss": 4.28773307800293,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.009512359276413918,
+ "learning_rate": 0.001920333624243228,
+ "loss": 4.266008377075195,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.009946667589247227,
+ "learning_rate": 0.001919051142812593,
+ "loss": 4.279062271118164,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.00899894442409277,
+ "learning_rate": 0.0019177683358452505,
+ "loss": 4.277679443359375,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.00909668579697609,
+ "learning_rate": 0.0019164852043747338,
+ "loss": 4.311397552490234,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.009787327609956264,
+ "learning_rate": 0.0019152017494348408,
+ "loss": 4.30875825881958,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.008983899839222431,
+ "learning_rate": 0.001913917972059626,
+ "loss": 4.259649276733398,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.008411582559347153,
+ "learning_rate": 0.0019126338732834071,
+ "loss": 4.259811878204346,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.009162918664515018,
+ "learning_rate": 0.0019113494541407579,
+ "loss": 4.271515846252441,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.011304561980068684,
+ "learning_rate": 0.0019100647156665125,
+ "loss": 4.269217491149902,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.012483394704759121,
+ "learning_rate": 0.0019087796588957617,
+ "loss": 4.2949299812316895,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.012091380544006824,
+ "learning_rate": 0.0019074942848638522,
+ "loss": 4.289752006530762,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.010011567734181881,
+ "learning_rate": 0.0019062085946063869,
+ "loss": 4.294095993041992,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.008766217157244682,
+ "learning_rate": 0.0019049225891592226,
+ "loss": 4.256425857543945,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.012862488627433777,
+ "learning_rate": 0.001903636269558471,
+ "loss": 4.287847995758057,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.013917063362896442,
+ "learning_rate": 0.0019023496368404972,
+ "loss": 4.282784938812256,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.012129023671150208,
+ "learning_rate": 0.0019010626920419173,
+ "loss": 4.2486419677734375,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.010348426178097725,
+ "learning_rate": 0.0018997754361995986,
+ "loss": 4.289836883544922,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.011821693740785122,
+ "learning_rate": 0.001898487870350662,
+ "loss": 4.214344024658203,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.013136724010109901,
+ "learning_rate": 0.0018971999955324742,
+ "loss": 4.24208402633667,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.013173168525099754,
+ "learning_rate": 0.0018959118127826538,
+ "loss": 4.270164489746094,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.011018159799277782,
+ "learning_rate": 0.0018946233231390656,
+ "loss": 4.2571916580200195,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.01018658559769392,
+ "learning_rate": 0.0018933345276398227,
+ "loss": 4.283267498016357,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.009412688203155994,
+ "learning_rate": 0.0018920454273232848,
+ "loss": 4.272198677062988,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.007799267303198576,
+ "learning_rate": 0.0018907560232280557,
+ "loss": 4.260125637054443,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.008747464045882225,
+ "learning_rate": 0.0018894663163929863,
+ "loss": 4.2539215087890625,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.008003434166312218,
+ "learning_rate": 0.0018881763078571689,
+ "loss": 4.245024681091309,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.007593846879899502,
+ "learning_rate": 0.0018868859986599415,
+ "loss": 4.237929344177246,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.008099420927464962,
+ "learning_rate": 0.0018855953898408815,
+ "loss": 4.2789106369018555,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.008270486257970333,
+ "learning_rate": 0.0018843044824398097,
+ "loss": 4.263522148132324,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.009260379709303379,
+ "learning_rate": 0.0018830132774967866,
+ "loss": 4.275835037231445,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.009289492852985859,
+ "learning_rate": 0.0018817217760521132,
+ "loss": 4.313778400421143,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.009493921883404255,
+ "learning_rate": 0.0018804299791463281,
+ "loss": 4.303672790527344,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.009020023047924042,
+ "learning_rate": 0.0018791378878202105,
+ "loss": 4.245473384857178,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.009446556679904461,
+ "learning_rate": 0.0018778455031147724,
+ "loss": 4.267327785491943,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.009499974548816681,
+ "learning_rate": 0.0018765528260712672,
+ "loss": 4.301205635070801,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.008399393409490585,
+ "learning_rate": 0.00187525985773118,
+ "loss": 4.307802677154541,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.009519701823592186,
+ "learning_rate": 0.0018739665991362327,
+ "loss": 4.284796714782715,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.01125232595950365,
+ "learning_rate": 0.0018726730513283804,
+ "loss": 4.285930633544922,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.011608537286520004,
+ "learning_rate": 0.00187137921534981,
+ "loss": 4.257689476013184,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.01064533181488514,
+ "learning_rate": 0.001870085092242944,
+ "loss": 4.249063968658447,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.009854599833488464,
+ "learning_rate": 0.0018687906830504321,
+ "loss": 4.255159378051758,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.01121739111840725,
+ "learning_rate": 0.0018674959888151576,
+ "loss": 4.287777900695801,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.5675368309021,
+ "eval_runtime": 51.8555,
+ "eval_samples_per_second": 47.092,
+ "eval_steps_per_second": 1.485,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.0104464590549469,
+ "learning_rate": 0.0018662010105802314,
+ "loss": 4.206080436706543,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.012800181284546852,
+ "learning_rate": 0.0018649057493889945,
+ "loss": 4.213550567626953,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.013382901437580585,
+ "learning_rate": 0.001863610206285015,
+ "loss": 4.190858840942383,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.014347050338983536,
+ "learning_rate": 0.0018623143823120894,
+ "loss": 4.22884464263916,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.01347152329981327,
+ "learning_rate": 0.001861018278514239,
+ "loss": 4.215579032897949,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.012786377221345901,
+ "learning_rate": 0.0018597218959357112,
+ "loss": 4.174999237060547,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.014137202873826027,
+ "learning_rate": 0.001858425235620979,
+ "loss": 4.21709680557251,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.013864399865269661,
+ "learning_rate": 0.001857128298614737,
+ "loss": 4.2375030517578125,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.010397658683359623,
+ "learning_rate": 0.0018558310859619048,
+ "loss": 4.231938362121582,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.00888819433748722,
+ "learning_rate": 0.0018545335987076223,
+ "loss": 4.195562839508057,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.010698067955672741,
+ "learning_rate": 0.001853235837897252,
+ "loss": 4.22847318649292,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.00987589918076992,
+ "learning_rate": 0.0018519378045763766,
+ "loss": 4.2147040367126465,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.011451782658696175,
+ "learning_rate": 0.001850639499790798,
+ "loss": 4.179263114929199,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.011224704794585705,
+ "learning_rate": 0.0018493409245865368,
+ "loss": 4.23030948638916,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.00996861606836319,
+ "learning_rate": 0.0018480420800098313,
+ "loss": 4.194283962249756,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.012143119238317013,
+ "learning_rate": 0.0018467429671071374,
+ "loss": 4.221704959869385,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.013126783072948456,
+ "learning_rate": 0.0018454435869251268,
+ "loss": 4.187748908996582,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.013031953014433384,
+ "learning_rate": 0.001844143940510687,
+ "loss": 4.248250961303711,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.012868993915617466,
+ "learning_rate": 0.001842844028910919,
+ "loss": 4.20161247253418,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.010102041065692902,
+ "learning_rate": 0.0018415438531731395,
+ "loss": 4.213654518127441,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.009599820710718632,
+ "learning_rate": 0.0018402434143448747,
+ "loss": 4.206313133239746,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.011536866426467896,
+ "learning_rate": 0.0018389427134738658,
+ "loss": 4.237072944641113,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.009558089077472687,
+ "learning_rate": 0.0018376417516080636,
+ "loss": 4.182195663452148,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01056219544261694,
+ "learning_rate": 0.0018363405297956302,
+ "loss": 4.232585906982422,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.012673584744334221,
+ "learning_rate": 0.0018350390490849354,
+ "loss": 4.228830337524414,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.011215884238481522,
+ "learning_rate": 0.0018337373105245598,
+ "loss": 4.216056823730469,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.011281915940344334,
+ "learning_rate": 0.00183243531516329,
+ "loss": 4.233404159545898,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.010381227359175682,
+ "learning_rate": 0.0018311330640501204,
+ "loss": 4.182013034820557,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.00900198519229889,
+ "learning_rate": 0.0018298305582342513,
+ "loss": 4.215488433837891,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.010198601521551609,
+ "learning_rate": 0.0018285277987650876,
+ "loss": 4.19389533996582,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.009965402074158192,
+ "learning_rate": 0.0018272247866922401,
+ "loss": 4.218005657196045,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.009299935773015022,
+ "learning_rate": 0.0018259215230655202,
+ "loss": 4.223684310913086,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.008977951481938362,
+ "learning_rate": 0.0018246180089349462,
+ "loss": 4.215778350830078,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.009261307306587696,
+ "learning_rate": 0.0018233142453507344,
+ "loss": 4.215066909790039,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.008322088047862053,
+ "learning_rate": 0.0018220102333633044,
+ "loss": 4.2188005447387695,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.008716144599020481,
+ "learning_rate": 0.0018207059740232739,
+ "loss": 4.187867641448975,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.008897020481526852,
+ "learning_rate": 0.0018194014683814625,
+ "loss": 4.218034267425537,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.007298174314200878,
+ "learning_rate": 0.0018180967174888857,
+ "loss": 4.2017011642456055,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.008491789922118187,
+ "learning_rate": 0.0018167917223967582,
+ "loss": 4.239873886108398,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.008413378149271011,
+ "learning_rate": 0.0018154864841564908,
+ "loss": 4.234771251678467,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.008141471073031425,
+ "learning_rate": 0.0018141810038196909,
+ "loss": 4.185944080352783,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.009480944834649563,
+ "learning_rate": 0.0018128752824381603,
+ "loss": 4.206586837768555,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.0106047997251153,
+ "learning_rate": 0.001811569321063894,
+ "loss": 4.235287666320801,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.010663514956831932,
+ "learning_rate": 0.0018102631207490838,
+ "loss": 4.204899311065674,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.011473523452877998,
+ "learning_rate": 0.001808956682546109,
+ "loss": 4.253668308258057,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01165992021560669,
+ "learning_rate": 0.0018076500075075456,
+ "loss": 4.209845542907715,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.011692410334944725,
+ "learning_rate": 0.0018063430966861564,
+ "loss": 4.250368595123291,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01061506848782301,
+ "learning_rate": 0.0018050359511348965,
+ "loss": 4.193821907043457,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.010510941036045551,
+ "learning_rate": 0.0018037285719069097,
+ "loss": 4.254327774047852,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.011587041430175304,
+ "learning_rate": 0.001802420960055528,
+ "loss": 4.2838945388793945,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.01257420051842928,
+ "learning_rate": 0.0018011131166342693,
+ "loss": 4.258241653442383,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.00936233438551426,
+ "learning_rate": 0.0017998050426968407,
+ "loss": 4.201854705810547,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.010093127377331257,
+ "learning_rate": 0.0017984967392971338,
+ "loss": 4.193051338195801,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.010032009333372116,
+ "learning_rate": 0.0017971882074892236,
+ "loss": 4.230212211608887,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.009455587714910507,
+ "learning_rate": 0.0017958794483273726,
+ "loss": 4.2157673835754395,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.009549238719046116,
+ "learning_rate": 0.001794570462866023,
+ "loss": 4.260272026062012,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.011059699580073357,
+ "learning_rate": 0.0017932612521598013,
+ "loss": 4.214818000793457,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.011653627268970013,
+ "learning_rate": 0.0017919518172635146,
+ "loss": 4.205112457275391,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.012438314035534859,
+ "learning_rate": 0.001790642159232151,
+ "loss": 4.179045677185059,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.011234673671424389,
+ "learning_rate": 0.0017893322791208784,
+ "loss": 4.239317417144775,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.010923845693469048,
+ "learning_rate": 0.0017880221779850434,
+ "loss": 4.230571269989014,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.009219309315085411,
+ "learning_rate": 0.001786711856880171,
+ "loss": 4.229068756103516,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.009849398396909237,
+ "learning_rate": 0.0017854013168619638,
+ "loss": 4.254772186279297,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.01117763202637434,
+ "learning_rate": 0.001784090558986299,
+ "loss": 4.252647399902344,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.010664612054824829,
+ "learning_rate": 0.0017827795843092316,
+ "loss": 4.204596519470215,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.00847950391471386,
+ "learning_rate": 0.0017814683938869895,
+ "loss": 4.211061954498291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.007305726408958435,
+ "learning_rate": 0.0017801569887759753,
+ "loss": 4.181221008300781,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.007970666512846947,
+ "learning_rate": 0.0017788453700327653,
+ "loss": 4.237855434417725,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.008873614482581615,
+ "learning_rate": 0.0017775335387141054,
+ "loss": 4.229188919067383,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.009691286832094193,
+ "learning_rate": 0.0017762214958769156,
+ "loss": 4.205491542816162,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.009360672906041145,
+ "learning_rate": 0.0017749092425782852,
+ "loss": 4.233466625213623,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.009895884431898594,
+ "learning_rate": 0.0017735967798754723,
+ "loss": 4.187763214111328,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.535672187805176,
+ "eval_runtime": 53.1736,
+ "eval_samples_per_second": 45.925,
+ "eval_steps_per_second": 1.448,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.00966129545122385,
+ "learning_rate": 0.0017722841088259045,
+ "loss": 4.1451735496521,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.010733380913734436,
+ "learning_rate": 0.0017709712304871774,
+ "loss": 4.1616363525390625,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.0117328567430377,
+ "learning_rate": 0.0017696581459170535,
+ "loss": 4.126331806182861,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.012640298344194889,
+ "learning_rate": 0.0017683448561734607,
+ "loss": 4.163730621337891,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.013074076734483242,
+ "learning_rate": 0.0017670313623144934,
+ "loss": 4.199502944946289,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.01132719125598669,
+ "learning_rate": 0.0017657176653984097,
+ "loss": 4.143380165100098,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.010008333250880241,
+ "learning_rate": 0.0017644037664836319,
+ "loss": 4.170320510864258,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.011090578511357307,
+ "learning_rate": 0.0017630896666287438,
+ "loss": 4.1571478843688965,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.011047456413507462,
+ "learning_rate": 0.0017617753668924924,
+ "loss": 4.101892948150635,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.010098653845489025,
+ "learning_rate": 0.0017604608683337857,
+ "loss": 4.154077529907227,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.010499860160052776,
+ "learning_rate": 0.00175914617201169,
+ "loss": 4.116868019104004,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.010939318686723709,
+ "learning_rate": 0.0017578312789854338,
+ "loss": 4.159144878387451,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.010622999630868435,
+ "learning_rate": 0.0017565161903144023,
+ "loss": 4.171411991119385,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.0108700692653656,
+ "learning_rate": 0.0017552009070581386,
+ "loss": 4.100748062133789,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.009794988669455051,
+ "learning_rate": 0.0017538854302763423,
+ "loss": 4.185161590576172,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.007750354241579771,
+ "learning_rate": 0.0017525697610288702,
+ "loss": 4.149083137512207,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.009923134930431843,
+ "learning_rate": 0.001751253900375732,
+ "loss": 4.205166816711426,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.010811141692101955,
+ "learning_rate": 0.0017499378493770936,
+ "loss": 4.131865978240967,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.010868688113987446,
+ "learning_rate": 0.0017486216090932733,
+ "loss": 4.148791790008545,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.011224126443266869,
+ "learning_rate": 0.001747305180584743,
+ "loss": 4.152503490447998,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.013262554071843624,
+ "learning_rate": 0.001745988564912124,
+ "loss": 4.164926528930664,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.013089599087834358,
+ "learning_rate": 0.001744671763136191,
+ "loss": 4.1626458168029785,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.01327646616846323,
+ "learning_rate": 0.0017433547763178665,
+ "loss": 4.189356803894043,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.014539042487740517,
+ "learning_rate": 0.0017420376055182233,
+ "loss": 4.176387786865234,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.013269098475575447,
+ "learning_rate": 0.001740720251798482,
+ "loss": 4.180716514587402,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.010922391898930073,
+ "learning_rate": 0.0017394027162200105,
+ "loss": 4.163928031921387,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.013355271890759468,
+ "learning_rate": 0.0017380849998443243,
+ "loss": 4.150012016296387,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.014149201102554798,
+ "learning_rate": 0.0017367671037330828,
+ "loss": 4.173713207244873,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01354879979044199,
+ "learning_rate": 0.0017354490289480918,
+ "loss": 4.150563716888428,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.013788704760372639,
+ "learning_rate": 0.0017341307765512998,
+ "loss": 4.213858604431152,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.011597222648561,
+ "learning_rate": 0.0017328123476047994,
+ "loss": 4.166989326477051,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.01055675558745861,
+ "learning_rate": 0.001731493743170824,
+ "loss": 4.172226428985596,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.011600397527217865,
+ "learning_rate": 0.0017301749643117506,
+ "loss": 4.1656951904296875,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.01114651933312416,
+ "learning_rate": 0.0017288560120900943,
+ "loss": 4.189825057983398,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.009811583906412125,
+ "learning_rate": 0.0017275368875685126,
+ "loss": 4.181164741516113,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.009758968837559223,
+ "learning_rate": 0.0017262175918097992,
+ "loss": 4.163188934326172,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.00999707542359829,
+ "learning_rate": 0.001724898125876887,
+ "loss": 4.184117317199707,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.009528180584311485,
+ "learning_rate": 0.001723578490832846,
+ "loss": 4.16407585144043,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.009717358276247978,
+ "learning_rate": 0.0017222586877408826,
+ "loss": 4.209892272949219,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.010909050703048706,
+ "learning_rate": 0.0017209387176643377,
+ "loss": 4.19977331161499,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.011234750039875507,
+ "learning_rate": 0.0017196185816666875,
+ "loss": 4.19363260269165,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.010300721041858196,
+ "learning_rate": 0.0017182982808115425,
+ "loss": 4.142427921295166,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.00941646471619606,
+ "learning_rate": 0.001716977816162644,
+ "loss": 4.159372329711914,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.009304054081439972,
+ "learning_rate": 0.0017156571887838679,
+ "loss": 4.207922458648682,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.009496180340647697,
+ "learning_rate": 0.0017143363997392184,
+ "loss": 4.137894630432129,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.009257436729967594,
+ "learning_rate": 0.0017130154500928326,
+ "loss": 4.198753833770752,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.00900586973875761,
+ "learning_rate": 0.0017116943409089746,
+ "loss": 4.160684585571289,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.009398494847118855,
+ "learning_rate": 0.00171037307325204,
+ "loss": 4.1819353103637695,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.008633013814687729,
+ "learning_rate": 0.0017090516481865483,
+ "loss": 4.176833152770996,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.00898425467312336,
+ "learning_rate": 0.0017077300667771484,
+ "loss": 4.190657138824463,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.008182398043572903,
+ "learning_rate": 0.0017064083300886155,
+ "loss": 4.173429489135742,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.009589143097400665,
+ "learning_rate": 0.0017050864391858478,
+ "loss": 4.1559014320373535,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.010304383002221584,
+ "learning_rate": 0.0017037643951338698,
+ "loss": 4.155453681945801,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.009943748824298382,
+ "learning_rate": 0.0017024421989978275,
+ "loss": 4.119734287261963,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.009238033555448055,
+ "learning_rate": 0.001701119851842992,
+ "loss": 4.188974857330322,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.008058690465986729,
+ "learning_rate": 0.0016997973547347531,
+ "loss": 4.19934606552124,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.008485953323543072,
+ "learning_rate": 0.0016984747087386233,
+ "loss": 4.165615558624268,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.008476605638861656,
+ "learning_rate": 0.0016971519149202347,
+ "loss": 4.21021842956543,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.008588708005845547,
+ "learning_rate": 0.0016958289743453385,
+ "loss": 4.208064556121826,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.009223178029060364,
+ "learning_rate": 0.0016945058880798037,
+ "loss": 4.158413887023926,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.009196167811751366,
+ "learning_rate": 0.0016931826571896166,
+ "loss": 4.203187465667725,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.009488921612501144,
+ "learning_rate": 0.001691859282740882,
+ "loss": 4.178227424621582,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.009253161028027534,
+ "learning_rate": 0.0016905357657998167,
+ "loss": 4.131230354309082,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.010069048963487148,
+ "learning_rate": 0.0016892121074327559,
+ "loss": 4.210734844207764,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.009945095516741276,
+ "learning_rate": 0.001687888308706147,
+ "loss": 4.184042930603027,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.010483640246093273,
+ "learning_rate": 0.0016865643706865505,
+ "loss": 4.156919479370117,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.011560715734958649,
+ "learning_rate": 0.0016852402944406391,
+ "loss": 4.185520172119141,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.010001277551054955,
+ "learning_rate": 0.0016839160810351978,
+ "loss": 4.226828575134277,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.010335934348404408,
+ "learning_rate": 0.0016825917315371202,
+ "loss": 4.183531284332275,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.011322776786983013,
+ "learning_rate": 0.001681267247013412,
+ "loss": 4.179635047912598,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.009663194417953491,
+ "learning_rate": 0.001679942628531186,
+ "loss": 4.165654182434082,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.010590246878564358,
+ "learning_rate": 0.0016786178771576643,
+ "loss": 4.1539459228515625,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.511386394500732,
+ "eval_runtime": 51.7751,
+ "eval_samples_per_second": 47.166,
+ "eval_steps_per_second": 1.487,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.01254354789853096,
+ "learning_rate": 0.0016772929939601728,
+ "loss": 4.104002475738525,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.01300987508147955,
+ "learning_rate": 0.0016759679800061483,
+ "loss": 4.127395153045654,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.013325816951692104,
+ "learning_rate": 0.0016746428363631299,
+ "loss": 4.131229400634766,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.014302870258688927,
+ "learning_rate": 0.001673317564098761,
+ "loss": 4.079358100891113,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.013776068575680256,
+ "learning_rate": 0.0016719921642807905,
+ "loss": 4.132065773010254,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.010788560844957829,
+ "learning_rate": 0.0016706666379770689,
+ "loss": 4.134626388549805,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.009766630828380585,
+ "learning_rate": 0.0016693409862555493,
+ "loss": 4.1645426750183105,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.011760029010474682,
+ "learning_rate": 0.0016680152101842842,
+ "loss": 4.110664367675781,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.011047836393117905,
+ "learning_rate": 0.0016666893108314288,
+ "loss": 4.129222869873047,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.009989139623939991,
+ "learning_rate": 0.0016653632892652344,
+ "loss": 4.076835632324219,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.010033783502876759,
+ "learning_rate": 0.001664037146554054,
+ "loss": 4.11555290222168,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.00996427796781063,
+ "learning_rate": 0.001662710883766337,
+ "loss": 4.131463527679443,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.01044023409485817,
+ "learning_rate": 0.0016613845019706288,
+ "loss": 4.098750114440918,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.010033872909843922,
+ "learning_rate": 0.0016600580022355712,
+ "loss": 4.161447048187256,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.009329567663371563,
+ "learning_rate": 0.0016587313856299012,
+ "loss": 4.146491527557373,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.010061589069664478,
+ "learning_rate": 0.0016574046532224504,
+ "loss": 4.107517719268799,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.009491910226643085,
+ "learning_rate": 0.0016560778060821424,
+ "loss": 4.098226547241211,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.008529345504939556,
+ "learning_rate": 0.0016547508452779943,
+ "loss": 4.1247711181640625,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.00936342403292656,
+ "learning_rate": 0.0016534237718791136,
+ "loss": 4.120573043823242,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.010255641303956509,
+ "learning_rate": 0.0016520965869547008,
+ "loss": 4.089067459106445,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.011638758704066277,
+ "learning_rate": 0.001650769291574044,
+ "loss": 4.074847221374512,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.014707857742905617,
+ "learning_rate": 0.0016494418868065214,
+ "loss": 4.153377532958984,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.01650111749768257,
+ "learning_rate": 0.0016481143737215986,
+ "loss": 4.134062767028809,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.014833372086286545,
+ "learning_rate": 0.00164678675338883,
+ "loss": 4.1344828605651855,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.012257886119186878,
+ "learning_rate": 0.0016454590268778538,
+ "loss": 4.098813533782959,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.013508322648704052,
+ "learning_rate": 0.0016441311952583959,
+ "loss": 4.106297492980957,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.01256900280714035,
+ "learning_rate": 0.001642803259600267,
+ "loss": 4.143631935119629,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.009885406121611595,
+ "learning_rate": 0.00164147522097336,
+ "loss": 4.120885848999023,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.012004674412310123,
+ "learning_rate": 0.0016401470804476518,
+ "loss": 4.089243412017822,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01224028691649437,
+ "learning_rate": 0.0016388188390932012,
+ "loss": 4.077765464782715,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.012511290609836578,
+ "learning_rate": 0.001637490497980149,
+ "loss": 4.161346435546875,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.010701716877520084,
+ "learning_rate": 0.0016361620581787143,
+ "loss": 4.10698938369751,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.010713113471865654,
+ "learning_rate": 0.001634833520759198,
+ "loss": 4.114283561706543,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.011018907651305199,
+ "learning_rate": 0.001633504886791978,
+ "loss": 4.134284019470215,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.011279584839940071,
+ "learning_rate": 0.0016321761573475117,
+ "loss": 4.1456804275512695,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.010642931796610355,
+ "learning_rate": 0.001630847333496332,
+ "loss": 4.160918235778809,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.011069678701460361,
+ "learning_rate": 0.0016295184163090482,
+ "loss": 4.136838912963867,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.011007166467607021,
+ "learning_rate": 0.0016281894068563443,
+ "loss": 4.129153728485107,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.010055703110992908,
+ "learning_rate": 0.0016268603062089802,
+ "loss": 4.138665676116943,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.009839230217039585,
+ "learning_rate": 0.0016255311154377875,
+ "loss": 4.1366496086120605,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.011191432364284992,
+ "learning_rate": 0.0016242018356136714,
+ "loss": 4.129880905151367,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.012130341492593288,
+ "learning_rate": 0.0016228724678076095,
+ "loss": 4.1425981521606445,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.012707795947790146,
+ "learning_rate": 0.0016215430130906482,
+ "loss": 4.115452766418457,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.012707809917628765,
+ "learning_rate": 0.001620213472533906,
+ "loss": 4.122758865356445,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.01182397548109293,
+ "learning_rate": 0.001618883847208569,
+ "loss": 4.140275955200195,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.010865340940654278,
+ "learning_rate": 0.001617554138185893,
+ "loss": 4.137759208679199,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.010973592288792133,
+ "learning_rate": 0.0016162243465371998,
+ "loss": 4.138947010040283,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.011247226968407631,
+ "learning_rate": 0.001614894473333879,
+ "loss": 4.115516662597656,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.010837487876415253,
+ "learning_rate": 0.0016135645196473853,
+ "loss": 4.160427093505859,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.009096396155655384,
+ "learning_rate": 0.0016122344865492385,
+ "loss": 4.1200032234191895,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.010671247728168964,
+ "learning_rate": 0.0016109043751110223,
+ "loss": 4.097919464111328,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.0107562355697155,
+ "learning_rate": 0.0016095741864043831,
+ "loss": 4.096290588378906,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.008909327909350395,
+ "learning_rate": 0.0016082439215010305,
+ "loss": 4.171017646789551,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.010305124334990978,
+ "learning_rate": 0.0016069135814727343,
+ "loss": 4.1450276374816895,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.011235613375902176,
+ "learning_rate": 0.0016055831673913255,
+ "loss": 4.134704113006592,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.00825632456690073,
+ "learning_rate": 0.0016042526803286949,
+ "loss": 4.107443332672119,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.009056378155946732,
+ "learning_rate": 0.0016029221213567921,
+ "loss": 4.132656097412109,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.008835164830088615,
+ "learning_rate": 0.001601591491547624,
+ "loss": 4.124729633331299,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.00915333442389965,
+ "learning_rate": 0.0016002607919732557,
+ "loss": 4.119500160217285,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.008532539941370487,
+ "learning_rate": 0.0015989300237058066,
+ "loss": 4.141661643981934,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.009952311404049397,
+ "learning_rate": 0.0015975991878174533,
+ "loss": 4.140929222106934,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.008920297026634216,
+ "learning_rate": 0.0015962682853804268,
+ "loss": 4.155877113342285,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.008770049549639225,
+ "learning_rate": 0.0015949373174670104,
+ "loss": 4.126394271850586,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.0078084105625748634,
+ "learning_rate": 0.0015936062851495411,
+ "loss": 4.128876209259033,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.00816641841083765,
+ "learning_rate": 0.0015922751895004078,
+ "loss": 4.137410640716553,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.008895459584891796,
+ "learning_rate": 0.0015909440315920505,
+ "loss": 4.07957649230957,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.008551528677344322,
+ "learning_rate": 0.001589612812496958,
+ "loss": 4.104213714599609,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.008325488306581974,
+ "learning_rate": 0.0015882815332876707,
+ "loss": 4.1584792137146,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.009341313503682613,
+ "learning_rate": 0.0015869501950367744,
+ "loss": 4.1417436599731445,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.00999050959944725,
+ "learning_rate": 0.0015856187988169067,
+ "loss": 4.136049270629883,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.008889932185411453,
+ "learning_rate": 0.0015842873457007484,
+ "loss": 4.16257381439209,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.009109573438763618,
+ "learning_rate": 0.0015829558367610268,
+ "loss": 4.171441555023193,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.485942840576172,
+ "eval_runtime": 52.1051,
+ "eval_samples_per_second": 46.867,
+ "eval_steps_per_second": 1.478,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.009881424717605114,
+ "learning_rate": 0.001581624273070515,
+ "loss": 4.059608459472656,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.011156721040606499,
+ "learning_rate": 0.0015802926557020294,
+ "loss": 4.047014236450195,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.012326992116868496,
+ "learning_rate": 0.0015789609857284309,
+ "loss": 4.06541109085083,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.01332208700478077,
+ "learning_rate": 0.0015776292642226213,
+ "loss": 4.042714595794678,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.012720128521323204,
+ "learning_rate": 0.0015762974922575446,
+ "loss": 4.098389625549316,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.010723215527832508,
+ "learning_rate": 0.0015749656709061854,
+ "loss": 4.059185981750488,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.010605281218886375,
+ "learning_rate": 0.001573633801241569,
+ "loss": 4.099538326263428,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.01101530622690916,
+ "learning_rate": 0.0015723018843367576,
+ "loss": 4.095674514770508,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.01135838869959116,
+ "learning_rate": 0.0015709699212648532,
+ "loss": 4.070535659790039,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.012119271792471409,
+ "learning_rate": 0.0015696379130989936,
+ "loss": 4.060605049133301,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.012783961370587349,
+ "learning_rate": 0.001568305860912355,
+ "loss": 4.075931072235107,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.011486220173537731,
+ "learning_rate": 0.0015669737657781461,
+ "loss": 4.075855255126953,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.01037552859634161,
+ "learning_rate": 0.0015656416287696136,
+ "loss": 4.087729454040527,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.009802743792533875,
+ "learning_rate": 0.0015643094509600354,
+ "loss": 4.0855607986450195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.010804438032209873,
+ "learning_rate": 0.001562977233422723,
+ "loss": 4.064528465270996,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.009505728259682655,
+ "learning_rate": 0.0015616449772310206,
+ "loss": 4.036775588989258,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.010634485632181168,
+ "learning_rate": 0.0015603126834583024,
+ "loss": 4.069923400878906,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.011392096057534218,
+ "learning_rate": 0.0015589803531779736,
+ "loss": 4.0812177658081055,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.011851722374558449,
+ "learning_rate": 0.0015576479874634683,
+ "loss": 4.09183406829834,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.011205444112420082,
+ "learning_rate": 0.0015563155873882502,
+ "loss": 4.07825231552124,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.009676349349319935,
+ "learning_rate": 0.0015549831540258093,
+ "loss": 4.052826404571533,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.009376165457069874,
+ "learning_rate": 0.0015536506884496638,
+ "loss": 4.02100133895874,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.00936955213546753,
+ "learning_rate": 0.0015523181917333564,
+ "loss": 4.049526214599609,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.00937784556299448,
+ "learning_rate": 0.0015509856649504564,
+ "loss": 4.096356391906738,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.010014008730649948,
+ "learning_rate": 0.0015496531091745555,
+ "loss": 4.03254508972168,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.009909634478390217,
+ "learning_rate": 0.0015483205254792705,
+ "loss": 4.067630767822266,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.009319501928985119,
+ "learning_rate": 0.0015469879149382399,
+ "loss": 4.094167709350586,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.008634721860289574,
+ "learning_rate": 0.0015456552786251244,
+ "loss": 4.122404098510742,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.009412454441189766,
+ "learning_rate": 0.0015443226176136049,
+ "loss": 4.081546783447266,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.010102108120918274,
+ "learning_rate": 0.0015429899329773814,
+ "loss": 4.049839019775391,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.010921654291450977,
+ "learning_rate": 0.0015416572257901746,
+ "loss": 4.097097396850586,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.010607713833451271,
+ "learning_rate": 0.0015403244971257226,
+ "loss": 4.084481239318848,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.011887960135936737,
+ "learning_rate": 0.0015389917480577806,
+ "loss": 4.074968338012695,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01089888159185648,
+ "learning_rate": 0.0015376589796601203,
+ "loss": 4.062132835388184,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.009934183210134506,
+ "learning_rate": 0.0015363261930065298,
+ "loss": 4.128092288970947,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.00878534372895956,
+ "learning_rate": 0.0015349933891708105,
+ "loss": 4.069051742553711,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.009928155690431595,
+ "learning_rate": 0.0015336605692267788,
+ "loss": 4.073912620544434,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.011091411113739014,
+ "learning_rate": 0.0015323277342482637,
+ "loss": 4.125944137573242,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.010314172133803368,
+ "learning_rate": 0.0015309948853091055,
+ "loss": 4.062620162963867,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.010423967614769936,
+ "learning_rate": 0.001529662023483158,
+ "loss": 4.101737976074219,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.009822997264564037,
+ "learning_rate": 0.0015283291498442817,
+ "loss": 4.064520359039307,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.009757095016539097,
+ "learning_rate": 0.001526996265466351,
+ "loss": 4.100624084472656,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.008650688454508781,
+ "learning_rate": 0.001525663371423246,
+ "loss": 4.109685897827148,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.007378234062343836,
+ "learning_rate": 0.0015243304687888553,
+ "loss": 4.044726848602295,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.008194336667656898,
+ "learning_rate": 0.001522997558637074,
+ "loss": 4.097923278808594,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.009764892049133778,
+ "learning_rate": 0.0015216646420418048,
+ "loss": 4.102148056030273,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.009242868050932884,
+ "learning_rate": 0.001520331720076953,
+ "loss": 4.068563461303711,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.010409324429929256,
+ "learning_rate": 0.001518998793816431,
+ "loss": 4.065944194793701,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.009215258061885834,
+ "learning_rate": 0.0015176658643341527,
+ "loss": 4.10223388671875,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.008065393194556236,
+ "learning_rate": 0.0015163329327040366,
+ "loss": 4.099432945251465,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.008498288691043854,
+ "learning_rate": 0.001515,
+ "loss": 4.109854698181152,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.008919150568544865,
+ "learning_rate": 0.0015136670672959633,
+ "loss": 4.128247261047363,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.009104429744184017,
+ "learning_rate": 0.0015123341356658474,
+ "loss": 4.068622589111328,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.008558365516364574,
+ "learning_rate": 0.0015110012061835692,
+ "loss": 4.114214897155762,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.007865076884627342,
+ "learning_rate": 0.0015096682799230477,
+ "loss": 4.11155366897583,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.007848534733057022,
+ "learning_rate": 0.0015083353579581958,
+ "loss": 4.045694351196289,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.008238030597567558,
+ "learning_rate": 0.0015070024413629264,
+ "loss": 4.093930721282959,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.007536398712545633,
+ "learning_rate": 0.0015056695312111448,
+ "loss": 4.106235504150391,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.008156892843544483,
+ "learning_rate": 0.001504336628576754,
+ "loss": 4.13060998916626,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.008144439198076725,
+ "learning_rate": 0.0015030037345336487,
+ "loss": 4.091304779052734,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.007757358253002167,
+ "learning_rate": 0.001501670850155718,
+ "loss": 4.069276809692383,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.00860623549669981,
+ "learning_rate": 0.0015003379765168425,
+ "loss": 4.107658386230469,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.008614457212388515,
+ "learning_rate": 0.0014990051146908949,
+ "loss": 4.106705665588379,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.009120592847466469,
+ "learning_rate": 0.001497672265751737,
+ "loss": 4.0891923904418945,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.009446980431675911,
+ "learning_rate": 0.0014963394307732213,
+ "loss": 4.098847389221191,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.010494766756892204,
+ "learning_rate": 0.0014950066108291894,
+ "loss": 4.115325450897217,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.010057658888399601,
+ "learning_rate": 0.0014936738069934706,
+ "loss": 4.099984169006348,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.009071771055459976,
+ "learning_rate": 0.0014923410203398796,
+ "loss": 4.091543674468994,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.009937075898051262,
+ "learning_rate": 0.00149100825194222,
+ "loss": 4.086492538452148,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.010222986340522766,
+ "learning_rate": 0.0014896755028742778,
+ "loss": 4.063841819763184,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.010513762012124062,
+ "learning_rate": 0.001488342774209826,
+ "loss": 4.11991548538208,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.009530222974717617,
+ "learning_rate": 0.0014870100670226192,
+ "loss": 4.043334007263184,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.467459678649902,
+ "eval_runtime": 54.1628,
+ "eval_samples_per_second": 45.086,
+ "eval_steps_per_second": 1.422,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.010274187661707401,
+ "learning_rate": 0.0014856773823863955,
+ "loss": 4.049561500549316,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.011448442004621029,
+ "learning_rate": 0.001484344721374876,
+ "loss": 4.019484996795654,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.011361192911863327,
+ "learning_rate": 0.0014830120850617599,
+ "loss": 4.0429887771606445,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.010250156745314598,
+ "learning_rate": 0.0014816794745207299,
+ "loss": 4.012682914733887,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.010024400427937508,
+ "learning_rate": 0.0014803468908254449,
+ "loss": 3.979409694671631,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.01153535582125187,
+ "learning_rate": 0.0014790143350495446,
+ "loss": 4.045780181884766,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.012632817961275578,
+ "learning_rate": 0.001477681808266644,
+ "loss": 4.057641983032227,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.013758293353021145,
+ "learning_rate": 0.0014763493115503366,
+ "loss": 4.061607360839844,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.014041583053767681,
+ "learning_rate": 0.001475016845974191,
+ "loss": 4.035029411315918,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.013629456982016563,
+ "learning_rate": 0.0014736844126117497,
+ "loss": 4.062169075012207,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.011023213155567646,
+ "learning_rate": 0.0014723520125365323,
+ "loss": 4.02731990814209,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.01001750584691763,
+ "learning_rate": 0.0014710196468220266,
+ "loss": 4.019882678985596,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.008321646600961685,
+ "learning_rate": 0.001469687316541698,
+ "loss": 4.005446434020996,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.008935946971178055,
+ "learning_rate": 0.0014683550227689796,
+ "loss": 4.017102241516113,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.009137130342423916,
+ "learning_rate": 0.001467022766577277,
+ "loss": 4.024102210998535,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.00805022194981575,
+ "learning_rate": 0.001465690549039965,
+ "loss": 3.9935250282287598,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.009204058907926083,
+ "learning_rate": 0.0014643583712303865,
+ "loss": 4.042818546295166,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.009812815114855766,
+ "learning_rate": 0.001463026234221854,
+ "loss": 4.0252485275268555,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.009955782443284988,
+ "learning_rate": 0.0014616941390876454,
+ "loss": 3.9839768409729004,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.009374645538628101,
+ "learning_rate": 0.001460362086901007,
+ "loss": 4.040245056152344,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.010008739307522774,
+ "learning_rate": 0.0014590300787351471,
+ "loss": 4.0306596755981445,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.009014200419187546,
+ "learning_rate": 0.001457698115663243,
+ "loss": 4.028713226318359,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.008381523191928864,
+ "learning_rate": 0.0014563661987584313,
+ "loss": 4.024718284606934,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.008627948351204395,
+ "learning_rate": 0.0014550343290938143,
+ "loss": 4.02303409576416,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.010257705114781857,
+ "learning_rate": 0.0014537025077424556,
+ "loss": 4.061058044433594,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.010759209282696247,
+ "learning_rate": 0.0014523707357773788,
+ "loss": 4.064718246459961,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.009966139681637287,
+ "learning_rate": 0.0014510390142715697,
+ "loss": 3.9919447898864746,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.00997584406286478,
+ "learning_rate": 0.001449707344297971,
+ "loss": 4.0431599617004395,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.010846944525837898,
+ "learning_rate": 0.0014483757269294858,
+ "loss": 4.0222697257995605,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.00991352554410696,
+ "learning_rate": 0.0014470441632389735,
+ "loss": 4.059438705444336,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.008994397707283497,
+ "learning_rate": 0.0014457126542992518,
+ "loss": 4.021494388580322,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.009567267261445522,
+ "learning_rate": 0.0014443812011830934,
+ "loss": 4.070810317993164,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.009199011139571667,
+ "learning_rate": 0.0014430498049632255,
+ "loss": 4.020938873291016,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.010440350510179996,
+ "learning_rate": 0.0014417184667123303,
+ "loss": 4.062529563903809,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.009962482377886772,
+ "learning_rate": 0.0014403871875030424,
+ "loss": 4.0687150955200195,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.009389379993081093,
+ "learning_rate": 0.0014390559684079502,
+ "loss": 4.055185794830322,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.011469941586256027,
+ "learning_rate": 0.0014377248104995924,
+ "loss": 4.041023254394531,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.011034650728106499,
+ "learning_rate": 0.0014363937148504588,
+ "loss": 4.054254531860352,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.010651151649653912,
+ "learning_rate": 0.0014350626825329897,
+ "loss": 4.037753582000732,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.010747930034995079,
+ "learning_rate": 0.0014337317146195733,
+ "loss": 4.059938907623291,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.00987576600164175,
+ "learning_rate": 0.0014324008121825468,
+ "loss": 4.075366973876953,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.008504371158778667,
+ "learning_rate": 0.0014310699762941936,
+ "loss": 4.026755332946777,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.009445052593946457,
+ "learning_rate": 0.001429739208026745,
+ "loss": 4.034431457519531,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.010560023598372936,
+ "learning_rate": 0.0014284085084523763,
+ "loss": 4.101158618927002,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.009920580312609673,
+ "learning_rate": 0.0014270778786432078,
+ "loss": 4.041321754455566,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.009854268282651901,
+ "learning_rate": 0.0014257473196713053,
+ "loss": 4.049882411956787,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.009825135581195354,
+ "learning_rate": 0.0014244168326086745,
+ "loss": 4.067440986633301,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.010506784543395042,
+ "learning_rate": 0.0014230864185272663,
+ "loss": 4.042476654052734,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.009527767077088356,
+ "learning_rate": 0.0014217560784989696,
+ "loss": 4.055248260498047,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.01013826485723257,
+ "learning_rate": 0.001420425813595617,
+ "loss": 4.05246639251709,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.011020111851394176,
+ "learning_rate": 0.0014190956248889778,
+ "loss": 4.032987117767334,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.010647117160260677,
+ "learning_rate": 0.0014177655134507614,
+ "loss": 4.093589782714844,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.009484780952334404,
+ "learning_rate": 0.001416435480352615,
+ "loss": 4.017437934875488,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.0114199910312891,
+ "learning_rate": 0.0014151055266661211,
+ "loss": 4.080643653869629,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.011137664318084717,
+ "learning_rate": 0.0014137756534628007,
+ "loss": 4.069380760192871,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.008972682990133762,
+ "learning_rate": 0.0014124458618141073,
+ "loss": 4.051383018493652,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.010311643593013287,
+ "learning_rate": 0.0014111161527914317,
+ "loss": 4.048953533172607,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.01113453321158886,
+ "learning_rate": 0.0014097865274660944,
+ "loss": 4.063544273376465,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.01091703586280346,
+ "learning_rate": 0.0014084569869093518,
+ "loss": 4.006721496582031,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.010672449134290218,
+ "learning_rate": 0.0014071275321923906,
+ "loss": 4.020220756530762,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.009682527743279934,
+ "learning_rate": 0.001405798164386328,
+ "loss": 4.057307243347168,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.009808676317334175,
+ "learning_rate": 0.0014044688845622126,
+ "loss": 4.105412483215332,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.010351189412176609,
+ "learning_rate": 0.0014031396937910202,
+ "loss": 4.066454887390137,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.010038275271654129,
+ "learning_rate": 0.001401810593143656,
+ "loss": 4.056752681732178,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01016094908118248,
+ "learning_rate": 0.001400481583690952,
+ "loss": 4.097933292388916,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.008683547377586365,
+ "learning_rate": 0.0013991526665036679,
+ "loss": 4.0386152267456055,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.009556332603096962,
+ "learning_rate": 0.0013978238426524882,
+ "loss": 4.05835485458374,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.009597044438123703,
+ "learning_rate": 0.001396495113208022,
+ "loss": 4.052363872528076,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.010021139867603779,
+ "learning_rate": 0.0013951664792408026,
+ "loss": 4.083431243896484,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.010611358098685741,
+ "learning_rate": 0.0013938379418212863,
+ "loss": 4.042330265045166,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.009132112376391888,
+ "learning_rate": 0.001392509502019852,
+ "loss": 4.058825969696045,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.009808709844946861,
+ "learning_rate": 0.0013911811609067992,
+ "loss": 4.093010902404785,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.4473161697387695,
+ "eval_runtime": 51.7738,
+ "eval_samples_per_second": 47.167,
+ "eval_steps_per_second": 1.487,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.011859052814543247,
+ "learning_rate": 0.0013898529195523488,
+ "loss": 3.9572651386260986,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.013210238888859749,
+ "learning_rate": 0.0013885247790266406,
+ "loss": 3.9812350273132324,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.012981324456632137,
+ "learning_rate": 0.001387196740399733,
+ "loss": 3.987725019454956,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.01424382347613573,
+ "learning_rate": 0.001385868804741604,
+ "loss": 3.9719889163970947,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.0154642965644598,
+ "learning_rate": 0.0013845409731221465,
+ "loss": 3.984248161315918,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.014779271557927132,
+ "learning_rate": 0.0013832132466111706,
+ "loss": 4.002912521362305,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.012921188026666641,
+ "learning_rate": 0.0013818856262784013,
+ "loss": 3.9588499069213867,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01165325753390789,
+ "learning_rate": 0.0013805581131934792,
+ "loss": 3.993288993835449,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.008908954448997974,
+ "learning_rate": 0.0013792307084259563,
+ "loss": 3.987151861190796,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.012040507979691029,
+ "learning_rate": 0.0013779034130452991,
+ "loss": 3.9675703048706055,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.013464349322021008,
+ "learning_rate": 0.0013765762281208868,
+ "loss": 3.9901437759399414,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.009626250714063644,
+ "learning_rate": 0.0013752491547220063,
+ "loss": 3.9373281002044678,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.00897257961332798,
+ "learning_rate": 0.0013739221939178582,
+ "loss": 4.00804328918457,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.010689414106309414,
+ "learning_rate": 0.00137259534677755,
+ "loss": 4.020804405212402,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.010462156496942043,
+ "learning_rate": 0.0013712686143700988,
+ "loss": 4.028973579406738,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.009135857224464417,
+ "learning_rate": 0.001369941997764429,
+ "loss": 3.9798216819763184,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.009298899210989475,
+ "learning_rate": 0.0013686154980293713,
+ "loss": 4.018845081329346,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.00979928020387888,
+ "learning_rate": 0.0013672891162336636,
+ "loss": 3.9946775436401367,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.008175542578101158,
+ "learning_rate": 0.0013659628534459457,
+ "loss": 3.981858491897583,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.008302642032504082,
+ "learning_rate": 0.0013646367107347662,
+ "loss": 3.991128444671631,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.008516009896993637,
+ "learning_rate": 0.0013633106891685718,
+ "loss": 4.004990100860596,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.007936482317745686,
+ "learning_rate": 0.0013619847898157164,
+ "loss": 3.995734214782715,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.008317600004374981,
+ "learning_rate": 0.001360659013744451,
+ "loss": 4.040653228759766,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.008481282740831375,
+ "learning_rate": 0.001359333362022931,
+ "loss": 3.969860553741455,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.008944948203861713,
+ "learning_rate": 0.0013580078357192096,
+ "loss": 3.9908909797668457,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.008897850289940834,
+ "learning_rate": 0.001356682435901239,
+ "loss": 4.04786491394043,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.009527456946671009,
+ "learning_rate": 0.001355357163636871,
+ "loss": 4.017423629760742,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.011714582331478596,
+ "learning_rate": 0.001354032019993852,
+ "loss": 3.9990692138671875,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.011083807796239853,
+ "learning_rate": 0.0013527070060398274,
+ "loss": 3.995265483856201,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.009878288954496384,
+ "learning_rate": 0.0013513821228423363,
+ "loss": 4.0364532470703125,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.009727329947054386,
+ "learning_rate": 0.0013500573714688136,
+ "loss": 4.035953521728516,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.009482033550739288,
+ "learning_rate": 0.0013487327529865879,
+ "loss": 4.015855312347412,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.009037062525749207,
+ "learning_rate": 0.0013474082684628798,
+ "loss": 4.024164199829102,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.009094092063605785,
+ "learning_rate": 0.0013460839189648028,
+ "loss": 4.037765979766846,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.008230654522776604,
+ "learning_rate": 0.0013447597055593612,
+ "loss": 4.006317615509033,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.008491289801895618,
+ "learning_rate": 0.0013434356293134501,
+ "loss": 4.042603969573975,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.008907693438231945,
+ "learning_rate": 0.0013421116912938533,
+ "loss": 4.008098602294922,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.008897818624973297,
+ "learning_rate": 0.0013407878925672438,
+ "loss": 3.9892945289611816,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.00957555789500475,
+ "learning_rate": 0.0013394642342001832,
+ "loss": 4.046009063720703,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.009948568418622017,
+ "learning_rate": 0.0013381407172591185,
+ "loss": 3.9760642051696777,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.01014955248683691,
+ "learning_rate": 0.0013368173428103835,
+ "loss": 4.023215293884277,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.008578895591199398,
+ "learning_rate": 0.0013354941119201966,
+ "loss": 4.002819061279297,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.008729875087738037,
+ "learning_rate": 0.001334171025654662,
+ "loss": 4.0340657234191895,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.009556280449032784,
+ "learning_rate": 0.0013328480850797655,
+ "loss": 3.983076333999634,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.009733820334076881,
+ "learning_rate": 0.0013315252912613766,
+ "loss": 4.066578388214111,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.009025252424180508,
+ "learning_rate": 0.0013302026452652474,
+ "loss": 3.9980697631835938,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.008485320024192333,
+ "learning_rate": 0.001328880148157008,
+ "loss": 4.018826961517334,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.009539696387946606,
+ "learning_rate": 0.0013275578010021726,
+ "loss": 4.011737823486328,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.010881016030907631,
+ "learning_rate": 0.0013262356048661305,
+ "loss": 4.007291793823242,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.010362236760556698,
+ "learning_rate": 0.0013249135608141523,
+ "loss": 3.9929394721984863,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.010296608321368694,
+ "learning_rate": 0.0013235916699113849,
+ "loss": 4.0015130043029785,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.011211174540221691,
+ "learning_rate": 0.0013222699332228517,
+ "loss": 4.005324363708496,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.009743994101881981,
+ "learning_rate": 0.0013209483518134525,
+ "loss": 4.0278143882751465,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.008563914336264133,
+ "learning_rate": 0.0013196269267479607,
+ "loss": 3.9997801780700684,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.009078902192413807,
+ "learning_rate": 0.0013183056590910255,
+ "loss": 4.026200294494629,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.009618895128369331,
+ "learning_rate": 0.0013169845499071675,
+ "loss": 4.027050018310547,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.008920644409954548,
+ "learning_rate": 0.0013156636002607822,
+ "loss": 4.000490665435791,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.008849202655255795,
+ "learning_rate": 0.0013143428112161323,
+ "loss": 4.044985294342041,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.009303621016442776,
+ "learning_rate": 0.001313022183837356,
+ "loss": 4.046281814575195,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.01076227892190218,
+ "learning_rate": 0.0013117017191884579,
+ "loss": 4.030576705932617,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.010400841943919659,
+ "learning_rate": 0.0013103814183333127,
+ "loss": 3.980081558227539,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.009459166787564754,
+ "learning_rate": 0.001309061282335663,
+ "loss": 4.014824867248535,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.009054969064891338,
+ "learning_rate": 0.001307741312259118,
+ "loss": 4.0485944747924805,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.009003991261124611,
+ "learning_rate": 0.0013064215091671545,
+ "loss": 3.9990320205688477,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.008959448896348476,
+ "learning_rate": 0.001305101874123113,
+ "loss": 4.038620948791504,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.009102243930101395,
+ "learning_rate": 0.0013037824081902015,
+ "loss": 4.030906677246094,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.008599337190389633,
+ "learning_rate": 0.0013024631124314877,
+ "loss": 4.02648401260376,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.008771977387368679,
+ "learning_rate": 0.0013011439879099056,
+ "loss": 4.027151107788086,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.009082595817744732,
+ "learning_rate": 0.00129982503568825,
+ "loss": 4.0194478034973145,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.009025354869663715,
+ "learning_rate": 0.0012985062568291763,
+ "loss": 4.044382095336914,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.00875102635473013,
+ "learning_rate": 0.0012971876523952014,
+ "loss": 4.0257568359375,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.009312704205513,
+ "learning_rate": 0.0012958692234487006,
+ "loss": 3.9964494705200195,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.431707859039307,
+ "eval_runtime": 50.7311,
+ "eval_samples_per_second": 48.136,
+ "eval_steps_per_second": 1.518,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.00992550328373909,
+ "learning_rate": 0.0012945509710519088,
+ "loss": 3.9253792762756348,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.009920453652739525,
+ "learning_rate": 0.0012932328962669173,
+ "loss": 3.9438819885253906,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.009437218308448792,
+ "learning_rate": 0.0012919150001556756,
+ "loss": 3.960963249206543,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.009616893716156483,
+ "learning_rate": 0.0012905972837799896,
+ "loss": 3.969301223754883,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.009107022546231747,
+ "learning_rate": 0.0012892797482015185,
+ "loss": 3.9626078605651855,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.011317167431116104,
+ "learning_rate": 0.0012879623944817772,
+ "loss": 3.946483850479126,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.010536770336329937,
+ "learning_rate": 0.001286645223682134,
+ "loss": 3.959566116333008,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.009349033236503601,
+ "learning_rate": 0.0012853282368638097,
+ "loss": 3.964435338973999,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.009950080886483192,
+ "learning_rate": 0.0012840114350878762,
+ "loss": 3.9729413986206055,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.010833023115992546,
+ "learning_rate": 0.001282694819415257,
+ "loss": 4.006421089172363,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.011538525111973286,
+ "learning_rate": 0.0012813783909067266,
+ "loss": 3.9558355808258057,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.010671315714716911,
+ "learning_rate": 0.0012800621506229065,
+ "loss": 4.0052337646484375,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.010835555382072926,
+ "learning_rate": 0.0012787460996242686,
+ "loss": 3.958641767501831,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.01082367543131113,
+ "learning_rate": 0.0012774302389711304,
+ "loss": 3.9818902015686035,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.009194346144795418,
+ "learning_rate": 0.0012761145697236583,
+ "loss": 3.9688854217529297,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.007859935984015465,
+ "learning_rate": 0.001274799092941862,
+ "loss": 3.9730639457702637,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.008534743450582027,
+ "learning_rate": 0.0012734838096855978,
+ "loss": 3.975003719329834,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.009369445964694023,
+ "learning_rate": 0.0012721687210145666,
+ "loss": 3.972907304763794,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.0090147340670228,
+ "learning_rate": 0.0012708538279883098,
+ "loss": 3.964256525039673,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.007745139300823212,
+ "learning_rate": 0.0012695391316662151,
+ "loss": 3.9504833221435547,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.008920101448893547,
+ "learning_rate": 0.0012682246331075075,
+ "loss": 3.9735097885131836,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.008987226523458958,
+ "learning_rate": 0.0012669103333712568,
+ "loss": 3.9482927322387695,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.00841289572417736,
+ "learning_rate": 0.0012655962335163685,
+ "loss": 3.965221881866455,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.008016116917133331,
+ "learning_rate": 0.0012642823346015902,
+ "loss": 3.9390206336975098,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.008470152504742146,
+ "learning_rate": 0.0012629686376855068,
+ "loss": 3.9257445335388184,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.008637234568595886,
+ "learning_rate": 0.0012616551438265394,
+ "loss": 3.9999356269836426,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.008036923594772816,
+ "learning_rate": 0.0012603418540829473,
+ "loss": 3.9755172729492188,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.008908843621611595,
+ "learning_rate": 0.0012590287695128232,
+ "loss": 3.9248545169830322,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.008743193000555038,
+ "learning_rate": 0.001257715891174096,
+ "loss": 3.9619390964508057,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.00834199320524931,
+ "learning_rate": 0.001256403220124528,
+ "loss": 3.961569309234619,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.009148893877863884,
+ "learning_rate": 0.001255090757421715,
+ "loss": 3.972104549407959,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.00925084762275219,
+ "learning_rate": 0.0012537785041230843,
+ "loss": 4.000775337219238,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.008659154176712036,
+ "learning_rate": 0.0012524664612858947,
+ "loss": 3.954564332962036,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.007787859998643398,
+ "learning_rate": 0.0012511546299672355,
+ "loss": 3.9498860836029053,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.00817918125540018,
+ "learning_rate": 0.0012498430112240246,
+ "loss": 4.003897666931152,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.009282548911869526,
+ "learning_rate": 0.001248531606113011,
+ "loss": 3.9621925354003906,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.009375610388815403,
+ "learning_rate": 0.001247220415690769,
+ "loss": 3.9977524280548096,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.009650378488004208,
+ "learning_rate": 0.001245909441013701,
+ "loss": 3.9882829189300537,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.009350919164717197,
+ "learning_rate": 0.0012445986831380366,
+ "loss": 4.019596099853516,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.008726303465664387,
+ "learning_rate": 0.0012432881431198289,
+ "loss": 3.9933621883392334,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.00829484686255455,
+ "learning_rate": 0.0012419778220149568,
+ "loss": 4.001237869262695,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.008744452148675919,
+ "learning_rate": 0.0012406677208791218,
+ "loss": 3.9323887825012207,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.008170492947101593,
+ "learning_rate": 0.0012393578407678496,
+ "loss": 3.925581455230713,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.007843422703444958,
+ "learning_rate": 0.001238048182736486,
+ "loss": 4.001023769378662,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.008914309553802013,
+ "learning_rate": 0.0012367387478401987,
+ "loss": 3.9804537296295166,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.009736751206219196,
+ "learning_rate": 0.0012354295371339774,
+ "loss": 3.9831767082214355,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.010061381384730339,
+ "learning_rate": 0.0012341205516726273,
+ "loss": 3.9462435245513916,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.010350692085921764,
+ "learning_rate": 0.0012328117925107765,
+ "loss": 4.012863636016846,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01039902213960886,
+ "learning_rate": 0.0012315032607028667,
+ "loss": 3.9811301231384277,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.011423371732234955,
+ "learning_rate": 0.0012301949573031595,
+ "loss": 3.9773542881011963,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.012449614703655243,
+ "learning_rate": 0.001228886883365731,
+ "loss": 3.969600200653076,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.010361429303884506,
+ "learning_rate": 0.0012275790399444726,
+ "loss": 4.035360336303711,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.009603082202374935,
+ "learning_rate": 0.0012262714280930904,
+ "loss": 3.9840846061706543,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.009183845482766628,
+ "learning_rate": 0.0012249640488651036,
+ "loss": 4.031542778015137,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.009694446809589863,
+ "learning_rate": 0.0012236569033138441,
+ "loss": 3.9935057163238525,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.0097816726192832,
+ "learning_rate": 0.0012223499924924545,
+ "loss": 3.9851176738739014,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.009413184598088264,
+ "learning_rate": 0.001221043317453891,
+ "loss": 3.9642205238342285,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.009960711002349854,
+ "learning_rate": 0.0012197368792509166,
+ "loss": 3.9373817443847656,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.010124670341610909,
+ "learning_rate": 0.0012184306789361059,
+ "loss": 3.958115577697754,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.011087656021118164,
+ "learning_rate": 0.00121712471756184,
+ "loss": 3.9607057571411133,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.009812800213694572,
+ "learning_rate": 0.001215818996180309,
+ "loss": 4.00722599029541,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.010305573232471943,
+ "learning_rate": 0.0012145135158435094,
+ "loss": 3.98728084564209,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.01019787136465311,
+ "learning_rate": 0.0012132082776032422,
+ "loss": 3.981172800064087,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.009260181337594986,
+ "learning_rate": 0.001211903282511115,
+ "loss": 4.034339904785156,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.00972998607903719,
+ "learning_rate": 0.001210598531618538,
+ "loss": 4.022885322570801,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.009853586554527283,
+ "learning_rate": 0.0012092940259767267,
+ "loss": 4.000235080718994,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.00961449183523655,
+ "learning_rate": 0.0012079897666366962,
+ "loss": 3.9951353073120117,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.010257176123559475,
+ "learning_rate": 0.0012066857546492657,
+ "loss": 3.9801371097564697,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.009443561546504498,
+ "learning_rate": 0.001205381991065054,
+ "loss": 3.9596502780914307,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.010389018803834915,
+ "learning_rate": 0.0012040784769344795,
+ "loss": 3.980527639389038,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.010322567075490952,
+ "learning_rate": 0.0012027752133077607,
+ "loss": 3.996258497238159,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.009987728670239449,
+ "learning_rate": 0.0012014722012349125,
+ "loss": 3.9493203163146973,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.421086311340332,
+ "eval_runtime": 52.5623,
+ "eval_samples_per_second": 46.459,
+ "eval_steps_per_second": 1.465,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.009405115619301796,
+ "learning_rate": 0.0012001694417657493,
+ "loss": 3.927506923675537,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.009974643588066101,
+ "learning_rate": 0.00119886693594988,
+ "loss": 3.9373550415039062,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.011331096291542053,
+ "learning_rate": 0.0011975646848367098,
+ "loss": 3.935594081878662,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.010538448579609394,
+ "learning_rate": 0.0011962626894754405,
+ "loss": 3.911579132080078,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.00989572424441576,
+ "learning_rate": 0.0011949609509150645,
+ "loss": 3.92525577545166,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.010533696971833706,
+ "learning_rate": 0.0011936594702043704,
+ "loss": 3.898075580596924,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.01051455456763506,
+ "learning_rate": 0.0011923582483919365,
+ "loss": 3.93463134765625,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.009662075899541378,
+ "learning_rate": 0.0011910572865261348,
+ "loss": 3.898221731185913,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.008874096907675266,
+ "learning_rate": 0.0011897565856551257,
+ "loss": 3.926787853240967,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.008126807399094105,
+ "learning_rate": 0.0011884561468268606,
+ "loss": 3.9158413410186768,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.007915553636848927,
+ "learning_rate": 0.001187155971089081,
+ "loss": 3.8891496658325195,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.008312922902405262,
+ "learning_rate": 0.0011858560594893129,
+ "loss": 3.932219982147217,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.008892328478395939,
+ "learning_rate": 0.0011845564130748734,
+ "loss": 3.8866336345672607,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.009187526069581509,
+ "learning_rate": 0.0011832570328928628,
+ "loss": 3.9464707374572754,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.008615494705736637,
+ "learning_rate": 0.0011819579199901693,
+ "loss": 3.9396629333496094,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.008649308234453201,
+ "learning_rate": 0.0011806590754134636,
+ "loss": 3.9325060844421387,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.008581082336604595,
+ "learning_rate": 0.0011793605002092021,
+ "loss": 3.9200944900512695,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.008502699434757233,
+ "learning_rate": 0.0011780621954236238,
+ "loss": 3.9656100273132324,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.0083576999604702,
+ "learning_rate": 0.0011767641621027477,
+ "loss": 3.9130215644836426,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.008196269161999226,
+ "learning_rate": 0.0011754664012923785,
+ "loss": 3.942369222640991,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.008857307955622673,
+ "learning_rate": 0.0011741689140380956,
+ "loss": 3.909585952758789,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.008935349993407726,
+ "learning_rate": 0.0011728717013852635,
+ "loss": 3.9248366355895996,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.008821200579404831,
+ "learning_rate": 0.0011715747643790211,
+ "loss": 3.9589288234710693,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.00861942209303379,
+ "learning_rate": 0.0011702781040642885,
+ "loss": 3.9381728172302246,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.009127574972808361,
+ "learning_rate": 0.0011689817214857613,
+ "loss": 3.897829532623291,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.008008269593119621,
+ "learning_rate": 0.0011676856176879107,
+ "loss": 3.930056095123291,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.010070208460092545,
+ "learning_rate": 0.0011663897937149853,
+ "loss": 3.892270803451538,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.00903113093227148,
+ "learning_rate": 0.001165094250611006,
+ "loss": 3.9766440391540527,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.008872415870428085,
+ "learning_rate": 0.0011637989894197692,
+ "loss": 3.9559712409973145,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.009524816647171974,
+ "learning_rate": 0.0011625040111848427,
+ "loss": 3.9607882499694824,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.009836779907345772,
+ "learning_rate": 0.0011612093169495678,
+ "loss": 3.9629592895507812,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.009586540050804615,
+ "learning_rate": 0.0011599149077570561,
+ "loss": 3.95208740234375,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.010495901107788086,
+ "learning_rate": 0.0011586207846501896,
+ "loss": 3.9439730644226074,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.010346266441047192,
+ "learning_rate": 0.0011573269486716202,
+ "loss": 3.959299087524414,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.010767634026706219,
+ "learning_rate": 0.0011560334008637674,
+ "loss": 3.900784969329834,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.012122200801968575,
+ "learning_rate": 0.0011547401422688202,
+ "loss": 3.9588193893432617,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.012914645485579967,
+ "learning_rate": 0.0011534471739287331,
+ "loss": 3.9278407096862793,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.011785333044826984,
+ "learning_rate": 0.0011521544968852273,
+ "loss": 3.992793560028076,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.008455781266093254,
+ "learning_rate": 0.0011508621121797899,
+ "loss": 3.96525239944458,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.009655282832682133,
+ "learning_rate": 0.0011495700208536714,
+ "loss": 3.9346890449523926,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.010537004098296165,
+ "learning_rate": 0.0011482782239478871,
+ "loss": 3.9534668922424316,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.009438651613891125,
+ "learning_rate": 0.0011469867225032135,
+ "loss": 3.983487606048584,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.010586646385490894,
+ "learning_rate": 0.0011456955175601908,
+ "loss": 3.939680576324463,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.009082718752324581,
+ "learning_rate": 0.001144404610159119,
+ "loss": 3.9558515548706055,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.009833602234721184,
+ "learning_rate": 0.0011431140013400591,
+ "loss": 3.9539828300476074,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.009142170660197735,
+ "learning_rate": 0.0011418236921428313,
+ "loss": 3.9442849159240723,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.008603853173553944,
+ "learning_rate": 0.0011405336836070137,
+ "loss": 3.9526913166046143,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.008838302455842495,
+ "learning_rate": 0.0011392439767719446,
+ "loss": 3.973104953765869,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.008867006748914719,
+ "learning_rate": 0.0011379545726767156,
+ "loss": 3.9832305908203125,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.00971697736531496,
+ "learning_rate": 0.001136665472360178,
+ "loss": 3.9563937187194824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.00957175437361002,
+ "learning_rate": 0.001135376676860935,
+ "loss": 3.975194215774536,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.008548660203814507,
+ "learning_rate": 0.001134088187217347,
+ "loss": 3.9839985370635986,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.008556334301829338,
+ "learning_rate": 0.0011328000044675264,
+ "loss": 3.9225516319274902,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.008440867997705936,
+ "learning_rate": 0.0011315121296493385,
+ "loss": 3.945028066635132,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.009643333032727242,
+ "learning_rate": 0.0011302245638004017,
+ "loss": 3.943120002746582,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.008832835592329502,
+ "learning_rate": 0.001128937307958083,
+ "loss": 3.9895920753479004,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.008720474317669868,
+ "learning_rate": 0.0011276503631595036,
+ "loss": 3.941950798034668,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.008099926635622978,
+ "learning_rate": 0.001126363730441529,
+ "loss": 3.9626870155334473,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.008525107055902481,
+ "learning_rate": 0.0011250774108407778,
+ "loss": 3.950535774230957,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.008998272009193897,
+ "learning_rate": 0.0011237914053936135,
+ "loss": 3.9460678100585938,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.009387437254190445,
+ "learning_rate": 0.001122505715136148,
+ "loss": 3.926604986190796,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.007804017513990402,
+ "learning_rate": 0.0011212203411042386,
+ "loss": 3.9894230365753174,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.008564616553485394,
+ "learning_rate": 0.0011199352843334876,
+ "loss": 3.9339959621429443,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.008726440370082855,
+ "learning_rate": 0.0011186505458592427,
+ "loss": 3.9867987632751465,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.008074100129306316,
+ "learning_rate": 0.0011173661267165932,
+ "loss": 3.9492292404174805,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.007688041776418686,
+ "learning_rate": 0.0011160820279403744,
+ "loss": 3.9425454139709473,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.007808993570506573,
+ "learning_rate": 0.0011147982505651595,
+ "loss": 3.935114860534668,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.009786820039153099,
+ "learning_rate": 0.0011135147956252659,
+ "loss": 3.959758758544922,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.009487155824899673,
+ "learning_rate": 0.00111223166415475,
+ "loss": 3.989048480987549,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.008407804183661938,
+ "learning_rate": 0.0011109488571874073,
+ "loss": 3.932347297668457,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.007158380467444658,
+ "learning_rate": 0.0011096663757567723,
+ "loss": 3.9545154571533203,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.008840581402182579,
+ "learning_rate": 0.001108384220896117,
+ "loss": 3.965984344482422,
+ "step": 2160
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 9.208840083996672e+17,
+ "train_batch_size": 16,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/baseline/checkpoint-2160/training_args.bin b/runs/baseline/checkpoint-2160/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7
--- /dev/null
+++ b/runs/baseline/checkpoint-2160/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97
+size 5265
diff --git a/runs/baseline/checkpoint-2520/chat_template.jinja b/runs/baseline/checkpoint-2520/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/baseline/checkpoint-2520/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/baseline/checkpoint-2520/config.json b/runs/baseline/checkpoint-2520/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/baseline/checkpoint-2520/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/baseline/checkpoint-2520/generation_config.json b/runs/baseline/checkpoint-2520/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/baseline/checkpoint-2520/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/baseline/checkpoint-2520/model.safetensors b/runs/baseline/checkpoint-2520/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..d37cd51501d9c0f34cbe72c8a6bb76b4a4c16c43
--- /dev/null
+++ b/runs/baseline/checkpoint-2520/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:18ed9230704ff7b89d1d7eb34a36f2e139ce85cc503b3719af861f86780c0ae5
+size 583356232
diff --git a/runs/baseline/checkpoint-2520/optimizer.pt b/runs/baseline/checkpoint-2520/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..e131ed4259dedbf462877c2a1e1e1890168c22a0
--- /dev/null
+++ b/runs/baseline/checkpoint-2520/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:58a72143ae108aa9b217d96413039e371b321c8f03f7a20e6f3d746c03825e73
+size 1166825803
diff --git a/runs/baseline/checkpoint-2520/rng_state_0.pth b/runs/baseline/checkpoint-2520/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..f6a767b5a760715c982d5f3b7abf7f7ae7024858
--- /dev/null
+++ b/runs/baseline/checkpoint-2520/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:77a1aaa8594be4aeac424d02802142f8409ed1e2897c1e46e3c37ed12b93e83c
+size 14917
diff --git a/runs/baseline/checkpoint-2520/rng_state_1.pth b/runs/baseline/checkpoint-2520/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..b0b2ce7324fa0e3d93e84f806452f72a932cb665
--- /dev/null
+++ b/runs/baseline/checkpoint-2520/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8be9b03fd0264df00f58b952952d0e7adf408ae412c760c45b6919b81dd07768
+size 14917
diff --git a/runs/baseline/checkpoint-2520/scheduler.pt b/runs/baseline/checkpoint-2520/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..74bd0e9c9ee7711fa33882e1d3ae8d7f5d1dead2
--- /dev/null
+++ b/runs/baseline/checkpoint-2520/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:37e4964cd4dada742240c68b526bbf0c58037d9596fb0e4b5651dac33cc25e70
+size 1465
diff --git a/runs/baseline/checkpoint-2520/tokenizer.json b/runs/baseline/checkpoint-2520/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/baseline/checkpoint-2520/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/baseline/checkpoint-2520/tokenizer_config.json b/runs/baseline/checkpoint-2520/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/baseline/checkpoint-2520/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/baseline/checkpoint-2520/trainer_state.json b/runs/baseline/checkpoint-2520/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..15978b8ede1af2855045a833f74c9d75e90d9c80
--- /dev/null
+++ b/runs/baseline/checkpoint-2520/trainer_state.json
@@ -0,0 +1,17946 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 35.0,
+ "eval_steps": 500,
+ "global_step": 2520,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.12165174633264542,
+ "learning_rate": 0.0,
+ "loss": 12.027063369750977,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.12109383195638657,
+ "learning_rate": 3e-05,
+ "loss": 12.028635025024414,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.12690064311027527,
+ "learning_rate": 6e-05,
+ "loss": 11.894524574279785,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.1537286937236786,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.672262191772461,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.11778244376182556,
+ "learning_rate": 0.00012,
+ "loss": 11.433671951293945,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.10874935984611511,
+ "learning_rate": 0.00015000000000000001,
+ "loss": 11.270309448242188,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.10641921311616898,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 11.155481338500977,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.1048506498336792,
+ "learning_rate": 0.00021000000000000004,
+ "loss": 11.066675186157227,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.10546834021806717,
+ "learning_rate": 0.00024,
+ "loss": 10.952295303344727,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.10614268481731415,
+ "learning_rate": 0.00027,
+ "loss": 10.811386108398438,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10529010742902756,
+ "learning_rate": 0.00030000000000000003,
+ "loss": 10.657712936401367,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.1051444336771965,
+ "learning_rate": 0.00033,
+ "loss": 10.47591781616211,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10420846194028854,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 10.280141830444336,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10323625057935715,
+ "learning_rate": 0.00039000000000000005,
+ "loss": 10.076366424560547,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10255401581525803,
+ "learning_rate": 0.00042000000000000007,
+ "loss": 9.856939315795898,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10151027888059616,
+ "learning_rate": 0.00045,
+ "loss": 9.632134437561035,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.1006326898932457,
+ "learning_rate": 0.00048,
+ "loss": 9.401637077331543,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.09820562601089478,
+ "learning_rate": 0.00051,
+ "loss": 9.177400588989258,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.0956626906991005,
+ "learning_rate": 0.00054,
+ "loss": 8.95518684387207,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.0922091156244278,
+ "learning_rate": 0.00057,
+ "loss": 8.75066089630127,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.08593932539224625,
+ "learning_rate": 0.0006000000000000001,
+ "loss": 8.557976722717285,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.08014766871929169,
+ "learning_rate": 0.00063,
+ "loss": 8.360604286193848,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.07381244748830795,
+ "learning_rate": 0.00066,
+ "loss": 8.210375785827637,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.06455052644014359,
+ "learning_rate": 0.0006900000000000001,
+ "loss": 8.07168960571289,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.054115474224090576,
+ "learning_rate": 0.0007199999999999999,
+ "loss": 7.960069179534912,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.043575264513492584,
+ "learning_rate": 0.00075,
+ "loss": 7.850685119628906,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.02976679615676403,
+ "learning_rate": 0.0007800000000000001,
+ "loss": 7.824377059936523,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.020552165806293488,
+ "learning_rate": 0.0008100000000000001,
+ "loss": 7.7695417404174805,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.018530743196606636,
+ "learning_rate": 0.0008400000000000001,
+ "loss": 7.782247543334961,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.02716820128262043,
+ "learning_rate": 0.00087,
+ "loss": 7.754945278167725,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.03638409823179245,
+ "learning_rate": 0.0009,
+ "loss": 7.799381256103516,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.041579652577638626,
+ "learning_rate": 0.00093,
+ "loss": 7.776629447937012,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.0304960235953331,
+ "learning_rate": 0.00096,
+ "loss": 7.749439239501953,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.053728822618722916,
+ "learning_rate": 0.00099,
+ "loss": 7.745059490203857,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.036322493106126785,
+ "learning_rate": 0.00102,
+ "loss": 7.752074718475342,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.03232959285378456,
+ "learning_rate": 0.00105,
+ "loss": 7.704257011413574,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.03246184438467026,
+ "learning_rate": 0.00108,
+ "loss": 7.667045593261719,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.036655545234680176,
+ "learning_rate": 0.00111,
+ "loss": 7.623445987701416,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.021042725071310997,
+ "learning_rate": 0.00114,
+ "loss": 7.582821846008301,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.022982975468039513,
+ "learning_rate": 0.00117,
+ "loss": 7.58036994934082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.024781864136457443,
+ "learning_rate": 0.0012000000000000001,
+ "loss": 7.5903000831604,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.03667089343070984,
+ "learning_rate": 0.00123,
+ "loss": 7.588233947753906,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.04083017632365227,
+ "learning_rate": 0.00126,
+ "loss": 7.548343658447266,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.02585385926067829,
+ "learning_rate": 0.00129,
+ "loss": 7.604833602905273,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.018793931230902672,
+ "learning_rate": 0.00132,
+ "loss": 7.56705379486084,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.025538936257362366,
+ "learning_rate": 0.00135,
+ "loss": 7.562107086181641,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.02819318138062954,
+ "learning_rate": 0.0013800000000000002,
+ "loss": 7.517213821411133,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.016239255666732788,
+ "learning_rate": 0.00141,
+ "loss": 7.4780755043029785,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.024470701813697815,
+ "learning_rate": 0.0014399999999999999,
+ "loss": 7.505887985229492,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.015499847941100597,
+ "learning_rate": 0.00147,
+ "loss": 7.484469890594482,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.017785200849175453,
+ "learning_rate": 0.0015,
+ "loss": 7.481822967529297,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.015952805057168007,
+ "learning_rate": 0.0015300000000000001,
+ "loss": 7.484804153442383,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.015771260485053062,
+ "learning_rate": 0.0015600000000000002,
+ "loss": 7.4495439529418945,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.024370530620217323,
+ "learning_rate": 0.00159,
+ "loss": 7.471208572387695,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.018173737451434135,
+ "learning_rate": 0.0016200000000000001,
+ "loss": 7.438447952270508,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.020564094185829163,
+ "learning_rate": 0.0016500000000000002,
+ "loss": 7.437922954559326,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.016154944896697998,
+ "learning_rate": 0.0016800000000000003,
+ "loss": 7.429378032684326,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.016817007213830948,
+ "learning_rate": 0.00171,
+ "loss": 7.406428337097168,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.03460947796702385,
+ "learning_rate": 0.00174,
+ "loss": 7.393331527709961,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.15410488843917847,
+ "learning_rate": 0.0017699999999999999,
+ "loss": 7.645856857299805,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.07443148642778397,
+ "learning_rate": 0.0018,
+ "loss": 7.549508094787598,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.0507572740316391,
+ "learning_rate": 0.00183,
+ "loss": 7.517921447753906,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.09682746231555939,
+ "learning_rate": 0.00186,
+ "loss": 7.543978691101074,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.2571517527103424,
+ "learning_rate": 0.00189,
+ "loss": 7.768949508666992,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.0317784883081913,
+ "learning_rate": 0.00192,
+ "loss": 7.509063720703125,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.0353970006108284,
+ "learning_rate": 0.0019500000000000001,
+ "loss": 7.522322654724121,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.03663994371891022,
+ "learning_rate": 0.00198,
+ "loss": 7.504723072052002,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.025481393560767174,
+ "learning_rate": 0.00201,
+ "loss": 7.452597141265869,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.03099915198981762,
+ "learning_rate": 0.00204,
+ "loss": 7.455441474914551,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.022921698167920113,
+ "learning_rate": 0.00207,
+ "loss": 7.4454827308654785,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.032388731837272644,
+ "learning_rate": 0.0021,
+ "loss": 7.467103958129883,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.022428609430789948,
+ "learning_rate": 0.00213,
+ "loss": 7.467540264129639,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.469719409942627,
+ "eval_runtime": 50.9714,
+ "eval_samples_per_second": 47.909,
+ "eval_steps_per_second": 1.511,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.017480911687016487,
+ "learning_rate": 0.00216,
+ "loss": 7.455810546875,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.016675131395459175,
+ "learning_rate": 0.00219,
+ "loss": 7.474848747253418,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.017450228333473206,
+ "learning_rate": 0.00222,
+ "loss": 7.442611217498779,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.016823675483465195,
+ "learning_rate": 0.0022500000000000003,
+ "loss": 7.454628944396973,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.016717543825507164,
+ "learning_rate": 0.00228,
+ "loss": 7.425317764282227,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.01644429750740528,
+ "learning_rate": 0.00231,
+ "loss": 7.4102864265441895,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.013104400597512722,
+ "learning_rate": 0.00234,
+ "loss": 7.4559173583984375,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.01175684668123722,
+ "learning_rate": 0.00237,
+ "loss": 7.438898086547852,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.013736684806644917,
+ "learning_rate": 0.0024000000000000002,
+ "loss": 7.449005603790283,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01176756713539362,
+ "learning_rate": 0.0024300000000000003,
+ "loss": 7.428421974182129,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.011860918253660202,
+ "learning_rate": 0.00246,
+ "loss": 7.411491394042969,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.010093477554619312,
+ "learning_rate": 0.00249,
+ "loss": 7.39311408996582,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.009453288279473782,
+ "learning_rate": 0.00252,
+ "loss": 7.387656211853027,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.010311692021787167,
+ "learning_rate": 0.00255,
+ "loss": 7.401076316833496,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.010053195990622044,
+ "learning_rate": 0.00258,
+ "loss": 7.39655876159668,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.009758037514984608,
+ "learning_rate": 0.00261,
+ "loss": 7.423122882843018,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008394946344196796,
+ "learning_rate": 0.00264,
+ "loss": 7.383601188659668,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.008457187563180923,
+ "learning_rate": 0.00267,
+ "loss": 7.400934219360352,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.010601334273815155,
+ "learning_rate": 0.0027,
+ "loss": 7.400241851806641,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.009009373374283314,
+ "learning_rate": 0.0027300000000000002,
+ "loss": 7.377348899841309,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.00737508200109005,
+ "learning_rate": 0.0027600000000000003,
+ "loss": 7.385881423950195,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.0075711836107075214,
+ "learning_rate": 0.0027900000000000004,
+ "loss": 7.348570823669434,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.00846901722252369,
+ "learning_rate": 0.00282,
+ "loss": 7.364287853240967,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.007160635199397802,
+ "learning_rate": 0.00285,
+ "loss": 7.3881635665893555,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.00682067358866334,
+ "learning_rate": 0.0028799999999999997,
+ "loss": 7.348663330078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.009108101949095726,
+ "learning_rate": 0.00291,
+ "loss": 7.356302261352539,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.015828927978873253,
+ "learning_rate": 0.00294,
+ "loss": 7.347576141357422,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.009683096781373024,
+ "learning_rate": 0.00297,
+ "loss": 7.323967933654785,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.010833270847797394,
+ "learning_rate": 0.003,
+ "loss": 7.3139495849609375,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.014905400574207306,
+ "learning_rate": 0.002999999401781161,
+ "loss": 7.302030086517334,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.023674853146076202,
+ "learning_rate": 0.002999997607125127,
+ "loss": 7.311811923980713,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.050460800528526306,
+ "learning_rate": 0.002999994616033342,
+ "loss": 7.41972541809082,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.02193174883723259,
+ "learning_rate": 0.002999990428508218,
+ "loss": 7.294468879699707,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.036261629313230515,
+ "learning_rate": 0.0029999850445531283,
+ "loss": 7.295568466186523,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.02734929881989956,
+ "learning_rate": 0.002999978464172409,
+ "loss": 7.309239864349365,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.017884107306599617,
+ "learning_rate": 0.002999970687371363,
+ "loss": 7.2427592277526855,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01700308918952942,
+ "learning_rate": 0.0029999617141562573,
+ "loss": 7.241730213165283,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.0209935512393713,
+ "learning_rate": 0.0029999515445343194,
+ "loss": 7.262085437774658,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.023207105696201324,
+ "learning_rate": 0.0029999401785137437,
+ "loss": 7.25450325012207,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.01907353848218918,
+ "learning_rate": 0.002999927616103687,
+ "loss": 7.23149299621582,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.016149666160345078,
+ "learning_rate": 0.0029999138573142715,
+ "loss": 7.210232734680176,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.017513928934931755,
+ "learning_rate": 0.002999898902156582,
+ "loss": 7.218451976776123,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.015002547763288021,
+ "learning_rate": 0.002999882750642667,
+ "loss": 7.169316291809082,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.014233081601560116,
+ "learning_rate": 0.0029998654027855406,
+ "loss": 7.194825172424316,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.01283260714262724,
+ "learning_rate": 0.0029998468585991783,
+ "loss": 7.15669059753418,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.014416799880564213,
+ "learning_rate": 0.0029998271180985223,
+ "loss": 7.143369674682617,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.014337058179080486,
+ "learning_rate": 0.0029998061812994757,
+ "loss": 7.127734184265137,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.016406210139393806,
+ "learning_rate": 0.002999784048218908,
+ "loss": 7.086504936218262,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.01680421642959118,
+ "learning_rate": 0.002999760718874652,
+ "loss": 7.089420318603516,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.010665548034012318,
+ "learning_rate": 0.0029997361932855016,
+ "loss": 7.102463722229004,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013479137793183327,
+ "learning_rate": 0.002999710471471218,
+ "loss": 7.079098701477051,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.012408486567437649,
+ "learning_rate": 0.0029996835534525246,
+ "loss": 7.0790114402771,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.012035196647047997,
+ "learning_rate": 0.0029996554392511094,
+ "loss": 7.058756351470947,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.00899985246360302,
+ "learning_rate": 0.0029996261288896225,
+ "loss": 7.0739240646362305,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.01072906143963337,
+ "learning_rate": 0.002999595622391679,
+ "loss": 7.085537910461426,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.01047342736274004,
+ "learning_rate": 0.0029995639197818577,
+ "loss": 7.060049057006836,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.02342618815600872,
+ "learning_rate": 0.0029995310210857006,
+ "loss": 7.039482116699219,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.011256859637796879,
+ "learning_rate": 0.0029994969263297134,
+ "loss": 7.040412902832031,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.008407019078731537,
+ "learning_rate": 0.0029994616355413666,
+ "loss": 7.0364274978637695,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.011006148532032967,
+ "learning_rate": 0.0029994251487490916,
+ "loss": 6.991468906402588,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.009103255346417427,
+ "learning_rate": 0.0029993874659822875,
+ "loss": 7.001059532165527,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.007875788025557995,
+ "learning_rate": 0.0029993485872713125,
+ "loss": 6.9974260330200195,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.007872611284255981,
+ "learning_rate": 0.0029993085126474913,
+ "loss": 6.980216979980469,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.010071957483887672,
+ "learning_rate": 0.0029992672421431115,
+ "loss": 6.97799015045166,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.01209618803113699,
+ "learning_rate": 0.0029992247757914237,
+ "loss": 6.97202205657959,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.014884858392179012,
+ "learning_rate": 0.0029991811136266425,
+ "loss": 6.976570129394531,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.014953463338315487,
+ "learning_rate": 0.0029991362556839453,
+ "loss": 6.971050262451172,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.008371063508093357,
+ "learning_rate": 0.002999090201999474,
+ "loss": 6.955556869506836,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.011926881037652493,
+ "learning_rate": 0.002999042952610332,
+ "loss": 6.943947792053223,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.013422893360257149,
+ "learning_rate": 0.002998994507554589,
+ "loss": 6.910022735595703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.009672224521636963,
+ "learning_rate": 0.002998944866871275,
+ "loss": 6.8994855880737305,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.012590581551194191,
+ "learning_rate": 0.0029988940306003853,
+ "loss": 6.88709020614624,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.945146083831787,
+ "eval_runtime": 51.5801,
+ "eval_samples_per_second": 47.344,
+ "eval_steps_per_second": 1.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.016554836183786392,
+ "learning_rate": 0.002998841998782877,
+ "loss": 6.916609764099121,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.011085660196840763,
+ "learning_rate": 0.0029987887714606723,
+ "loss": 6.884910583496094,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.008596168830990791,
+ "learning_rate": 0.0029987343486766547,
+ "loss": 6.900066375732422,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.014487302862107754,
+ "learning_rate": 0.0029986787304746715,
+ "loss": 6.8693342208862305,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.015636997297406197,
+ "learning_rate": 0.002998621916899534,
+ "loss": 6.903670787811279,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.013275462202727795,
+ "learning_rate": 0.0029985639079970153,
+ "loss": 6.889361381530762,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.009044744074344635,
+ "learning_rate": 0.002998504703813852,
+ "loss": 6.8549981117248535,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.011327696964144707,
+ "learning_rate": 0.002998444304397745,
+ "loss": 6.856151103973389,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.01310331467539072,
+ "learning_rate": 0.0029983827097973554,
+ "loss": 6.834944725036621,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.010003753937780857,
+ "learning_rate": 0.00299831992006231,
+ "loss": 6.825613975524902,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.009928115643560886,
+ "learning_rate": 0.0029982559352431967,
+ "loss": 6.828935146331787,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.012302150949835777,
+ "learning_rate": 0.0029981907553915674,
+ "loss": 6.821207046508789,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.020555168390274048,
+ "learning_rate": 0.002998124380559936,
+ "loss": 6.8184895515441895,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.011417638510465622,
+ "learning_rate": 0.0029980568108017797,
+ "loss": 6.788754463195801,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.010772978886961937,
+ "learning_rate": 0.002997988046171538,
+ "loss": 6.7861433029174805,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.012753859162330627,
+ "learning_rate": 0.002997918086724614,
+ "loss": 6.788776397705078,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.011327114887535572,
+ "learning_rate": 0.0029978469325173717,
+ "loss": 6.779160022735596,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01235753484070301,
+ "learning_rate": 0.002997774583607139,
+ "loss": 6.799554824829102,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.01417156495153904,
+ "learning_rate": 0.0029977010400522074,
+ "loss": 6.77957820892334,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.021321553736925125,
+ "learning_rate": 0.002997626301911828,
+ "loss": 6.785632133483887,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.03614472970366478,
+ "learning_rate": 0.002997550369246217,
+ "loss": 6.836414337158203,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03934052586555481,
+ "learning_rate": 0.002997473242116551,
+ "loss": 6.854659557342529,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.016240060329437256,
+ "learning_rate": 0.0029973949205849707,
+ "loss": 6.789690971374512,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.031164905056357384,
+ "learning_rate": 0.0029973154047145782,
+ "loss": 6.8363776206970215,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.013051844201982021,
+ "learning_rate": 0.002997234694569438,
+ "loss": 6.803576469421387,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.022611448541283607,
+ "learning_rate": 0.002997152790214577,
+ "loss": 6.762107849121094,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.017812907695770264,
+ "learning_rate": 0.0029970696917159833,
+ "loss": 6.776218414306641,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.015453332103788853,
+ "learning_rate": 0.0029969853991406088,
+ "loss": 6.742618083953857,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.020630091428756714,
+ "learning_rate": 0.0029968999125563655,
+ "loss": 6.765475749969482,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.014672341756522655,
+ "learning_rate": 0.0029968132320321297,
+ "loss": 6.754009246826172,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.016540220007300377,
+ "learning_rate": 0.0029967253576377376,
+ "loss": 6.734752655029297,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.013495649211108685,
+ "learning_rate": 0.0029966362894439874,
+ "loss": 6.752676010131836,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015936989337205887,
+ "learning_rate": 0.0029965460275226405,
+ "loss": 6.725708961486816,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010431243106722832,
+ "learning_rate": 0.0029964545719464194,
+ "loss": 6.698376655578613,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01669321395456791,
+ "learning_rate": 0.0029963619227890083,
+ "loss": 6.711770534515381,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.013308997265994549,
+ "learning_rate": 0.0029962680801250514,
+ "loss": 6.705286979675293,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011648476123809814,
+ "learning_rate": 0.002996173044030159,
+ "loss": 6.690052509307861,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.03942161798477173,
+ "learning_rate": 0.0029960768145808965,
+ "loss": 6.711885452270508,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.01571274921298027,
+ "learning_rate": 0.002995979391854797,
+ "loss": 6.715847969055176,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.00829714722931385,
+ "learning_rate": 0.0029958807759303515,
+ "loss": 6.685308933258057,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.01374727487564087,
+ "learning_rate": 0.0029957809668870114,
+ "loss": 6.695744514465332,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.010857059620320797,
+ "learning_rate": 0.0029956799648051934,
+ "loss": 6.658294677734375,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.008897246792912483,
+ "learning_rate": 0.002995577769766272,
+ "loss": 6.642439842224121,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015880772843956947,
+ "learning_rate": 0.002995474381852584,
+ "loss": 6.674907684326172,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.021623898297548294,
+ "learning_rate": 0.0029953698011474265,
+ "loss": 6.66236686706543,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.03084627166390419,
+ "learning_rate": 0.0029952640277350596,
+ "loss": 6.666715621948242,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.040561504662036896,
+ "learning_rate": 0.002995157061700702,
+ "loss": 6.649062156677246,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.031126515939831734,
+ "learning_rate": 0.0029950489031305345,
+ "loss": 6.640542030334473,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.00872360821813345,
+ "learning_rate": 0.0029949395521116987,
+ "loss": 6.635221481323242,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.023432321846485138,
+ "learning_rate": 0.0029948290087322966,
+ "loss": 6.642196178436279,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.021096013486385345,
+ "learning_rate": 0.0029947172730813913,
+ "loss": 6.643655300140381,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.014494858682155609,
+ "learning_rate": 0.002994604345249006,
+ "loss": 6.612720966339111,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.05225599184632301,
+ "learning_rate": 0.0029944902253261248,
+ "loss": 6.60659646987915,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.014311269856989384,
+ "learning_rate": 0.0029943749134046915,
+ "loss": 6.54373025894165,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.010777472518384457,
+ "learning_rate": 0.0029942584095776115,
+ "loss": 6.599837303161621,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.012498079799115658,
+ "learning_rate": 0.002994140713938749,
+ "loss": 6.589657783508301,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.015775132924318314,
+ "learning_rate": 0.00299402182658293,
+ "loss": 6.577511787414551,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.017978230491280556,
+ "learning_rate": 0.0029939017476059396,
+ "loss": 6.58717155456543,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.026807745918631554,
+ "learning_rate": 0.0029937804771045236,
+ "loss": 6.578145503997803,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.04853282868862152,
+ "learning_rate": 0.0029936580151763872,
+ "loss": 6.609755516052246,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.1057916209101677,
+ "learning_rate": 0.002993534361920195,
+ "loss": 6.870367050170898,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.04925911873579025,
+ "learning_rate": 0.002993409517435574,
+ "loss": 6.871556282043457,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.05288556218147278,
+ "learning_rate": 0.0029932834818231077,
+ "loss": 6.808315753936768,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.03831710293889046,
+ "learning_rate": 0.0029931562551843405,
+ "loss": 6.827617645263672,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.03665117174386978,
+ "learning_rate": 0.002993027837621778,
+ "loss": 6.767284870147705,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.026216736063361168,
+ "learning_rate": 0.0029928982292388826,
+ "loss": 6.733659744262695,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.024796854704618454,
+ "learning_rate": 0.002992767430140078,
+ "loss": 6.759577751159668,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.02410992607474327,
+ "learning_rate": 0.002992635440430747,
+ "loss": 6.701825141906738,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.019924134016036987,
+ "learning_rate": 0.0029925022602172308,
+ "loss": 6.677121162414551,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.022516867145895958,
+ "learning_rate": 0.00299236788960683,
+ "loss": 6.668622970581055,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.025068556889891624,
+ "learning_rate": 0.002992232328707806,
+ "loss": 6.67718505859375,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.02574208751320839,
+ "learning_rate": 0.002992095577629376,
+ "loss": 6.679594039916992,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.657555103302002,
+ "eval_runtime": 51.0199,
+ "eval_samples_per_second": 47.864,
+ "eval_steps_per_second": 1.509,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.017298126593232155,
+ "learning_rate": 0.0029919576364817184,
+ "loss": 6.620774269104004,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.018840325996279716,
+ "learning_rate": 0.002991818505375971,
+ "loss": 6.613316535949707,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.020813113078475,
+ "learning_rate": 0.002991678184424228,
+ "loss": 6.621555328369141,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.026581479236483574,
+ "learning_rate": 0.0029915366737395437,
+ "loss": 6.613471031188965,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.02084030583500862,
+ "learning_rate": 0.0029913939734359306,
+ "loss": 6.559704780578613,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.01327502354979515,
+ "learning_rate": 0.00299125008362836,
+ "loss": 6.564910888671875,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.015091734007000923,
+ "learning_rate": 0.0029911050044327614,
+ "loss": 6.571262836456299,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.015463982708752155,
+ "learning_rate": 0.002990958735966023,
+ "loss": 6.565750598907471,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.013636826537549496,
+ "learning_rate": 0.0029908112783459894,
+ "loss": 6.57981014251709,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.010613273829221725,
+ "learning_rate": 0.0029906626316914655,
+ "loss": 6.538505554199219,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.01804237812757492,
+ "learning_rate": 0.0029905127961222127,
+ "loss": 6.5282511711120605,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.015131509862840176,
+ "learning_rate": 0.0029903617717589506,
+ "loss": 6.531335830688477,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.010518206283450127,
+ "learning_rate": 0.002990209558723358,
+ "loss": 6.521886825561523,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.011360022239387035,
+ "learning_rate": 0.002990056157138069,
+ "loss": 6.528301239013672,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.011556181125342846,
+ "learning_rate": 0.0029899015671266773,
+ "loss": 6.497917652130127,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.01365791168063879,
+ "learning_rate": 0.0029897457888137325,
+ "loss": 6.52849006652832,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.01091043371707201,
+ "learning_rate": 0.0029895888223247426,
+ "loss": 6.492197036743164,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.008484847843647003,
+ "learning_rate": 0.002989430667786173,
+ "loss": 6.497377395629883,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.00836879201233387,
+ "learning_rate": 0.0029892713253254466,
+ "loss": 6.495189666748047,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.009781497530639172,
+ "learning_rate": 0.0029891107950709406,
+ "loss": 6.454085826873779,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.009603648446500301,
+ "learning_rate": 0.0029889490771519934,
+ "loss": 6.499475479125977,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.007991855964064598,
+ "learning_rate": 0.0029887861716988975,
+ "loss": 6.460785865783691,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.008447187952697277,
+ "learning_rate": 0.0029886220788429025,
+ "loss": 6.504582405090332,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.010945498012006283,
+ "learning_rate": 0.0029884567987162156,
+ "loss": 6.49589729309082,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.011096215806901455,
+ "learning_rate": 0.0029882903314519997,
+ "loss": 6.460657119750977,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.013094739057123661,
+ "learning_rate": 0.0029881226771843747,
+ "loss": 6.462450981140137,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.015423037111759186,
+ "learning_rate": 0.0029879538360484165,
+ "loss": 6.45565128326416,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.0182222630828619,
+ "learning_rate": 0.002987783808180157,
+ "loss": 6.452227592468262,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.01981227844953537,
+ "learning_rate": 0.002987612593716585,
+ "loss": 6.477701663970947,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.026733700186014175,
+ "learning_rate": 0.002987440192795645,
+ "loss": 6.452212333679199,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.03504155948758125,
+ "learning_rate": 0.0029872666055562372,
+ "loss": 6.509665489196777,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.027970125898718834,
+ "learning_rate": 0.002987091832138217,
+ "loss": 6.487089157104492,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.013714591972529888,
+ "learning_rate": 0.0029869158726823974,
+ "loss": 6.469569206237793,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.016472695395350456,
+ "learning_rate": 0.0029867387273305453,
+ "loss": 6.461688995361328,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.018457358703017235,
+ "learning_rate": 0.0029865603962253833,
+ "loss": 6.490985870361328,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.011045100167393684,
+ "learning_rate": 0.0029863808795105896,
+ "loss": 6.403701305389404,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.014126897789537907,
+ "learning_rate": 0.0029862001773307972,
+ "loss": 6.379804611206055,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.014015335589647293,
+ "learning_rate": 0.0029860182898315954,
+ "loss": 6.404377460479736,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.010032658465206623,
+ "learning_rate": 0.002985835217159527,
+ "loss": 6.40291690826416,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.015469838865101337,
+ "learning_rate": 0.0029856509594620912,
+ "loss": 6.4414520263671875,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.011106910184025764,
+ "learning_rate": 0.0029854655168877397,
+ "loss": 6.396444320678711,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.01262320950627327,
+ "learning_rate": 0.0029852788895858816,
+ "loss": 6.385222434997559,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.012783976271748543,
+ "learning_rate": 0.002985091077706878,
+ "loss": 6.3818135261535645,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.01224067248404026,
+ "learning_rate": 0.0029849020814020468,
+ "loss": 6.37886905670166,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.010540086776018143,
+ "learning_rate": 0.0029847119008236576,
+ "loss": 6.37799072265625,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.012479913420975208,
+ "learning_rate": 0.0029845205361249366,
+ "loss": 6.365094184875488,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.010080077685415745,
+ "learning_rate": 0.0029843279874600622,
+ "loss": 6.371852874755859,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.013266298919916153,
+ "learning_rate": 0.002984134254984168,
+ "loss": 6.380039215087891,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.01014771405607462,
+ "learning_rate": 0.002983939338853339,
+ "loss": 6.370337963104248,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.008164579048752785,
+ "learning_rate": 0.0029837432392246185,
+ "loss": 6.331899642944336,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.01020442508161068,
+ "learning_rate": 0.002983545956255998,
+ "loss": 6.35077428817749,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.009898579679429531,
+ "learning_rate": 0.002983347490106426,
+ "loss": 6.371985912322998,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.009668919257819653,
+ "learning_rate": 0.002983147840935802,
+ "loss": 6.356776237487793,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.012560366652905941,
+ "learning_rate": 0.0029829470089049815,
+ "loss": 6.357911109924316,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.018435249105095863,
+ "learning_rate": 0.0029827449941757698,
+ "loss": 6.366821765899658,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.029046855866909027,
+ "learning_rate": 0.002982541796910927,
+ "loss": 6.367595672607422,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.043010637164115906,
+ "learning_rate": 0.002982337417274165,
+ "loss": 6.413588523864746,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.045470818877220154,
+ "learning_rate": 0.0029821318554301497,
+ "loss": 6.435015678405762,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.02148726023733616,
+ "learning_rate": 0.002981925111544498,
+ "loss": 6.330894470214844,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.026808688417077065,
+ "learning_rate": 0.00298171718578378,
+ "loss": 6.384839057922363,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.018495213240385056,
+ "learning_rate": 0.002981508078315517,
+ "loss": 6.36277961730957,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.025393597781658173,
+ "learning_rate": 0.0029812977893081834,
+ "loss": 6.3702239990234375,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020856289193034172,
+ "learning_rate": 0.002981086318931206,
+ "loss": 6.354547500610352,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.017827816307544708,
+ "learning_rate": 0.002980873667354962,
+ "loss": 6.339264869689941,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.02639714442193508,
+ "learning_rate": 0.002980659834750781,
+ "loss": 6.3254194259643555,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.015238096937537193,
+ "learning_rate": 0.002980444821290944,
+ "loss": 6.365842819213867,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.012863729149103165,
+ "learning_rate": 0.002980228627148684,
+ "loss": 6.356776714324951,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.014739899896085262,
+ "learning_rate": 0.0029800112524981845,
+ "loss": 6.332481384277344,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.013582171872258186,
+ "learning_rate": 0.0029797926975145797,
+ "loss": 6.311968803405762,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.012704637832939625,
+ "learning_rate": 0.0029795729623739555,
+ "loss": 6.321654319763184,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.012550560757517815,
+ "learning_rate": 0.0029793520472533495,
+ "loss": 6.298409461975098,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.011170165613293648,
+ "learning_rate": 0.002979129952330748,
+ "loss": 6.294830799102783,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.326757907867432,
+ "eval_runtime": 51.6566,
+ "eval_samples_per_second": 47.274,
+ "eval_steps_per_second": 1.491,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.010947446338832378,
+ "learning_rate": 0.0029789066777850897,
+ "loss": 6.286581039428711,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.010097995400428772,
+ "learning_rate": 0.002978682223796262,
+ "loss": 6.288301467895508,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.01224475260823965,
+ "learning_rate": 0.002978456590545104,
+ "loss": 6.264986991882324,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.009734537452459335,
+ "learning_rate": 0.0029782297782134045,
+ "loss": 6.304243087768555,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.01136956550180912,
+ "learning_rate": 0.0029780017869839013,
+ "loss": 6.244938373565674,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.00935293547809124,
+ "learning_rate": 0.002977772617040284,
+ "loss": 6.251852035522461,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.011118784546852112,
+ "learning_rate": 0.0029775422685671894,
+ "loss": 6.256633758544922,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.010837923735380173,
+ "learning_rate": 0.0029773107417502055,
+ "loss": 6.263411045074463,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.011920305900275707,
+ "learning_rate": 0.0029770780367758703,
+ "loss": 6.239311218261719,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.014117483049631119,
+ "learning_rate": 0.0029768441538316685,
+ "loss": 6.253314018249512,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.022502148523926735,
+ "learning_rate": 0.002976609093106036,
+ "loss": 6.246156692504883,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.03348720073699951,
+ "learning_rate": 0.002976372854788357,
+ "loss": 6.255914688110352,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.03427121415734291,
+ "learning_rate": 0.002976135439068965,
+ "loss": 6.289946556091309,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.02407355234026909,
+ "learning_rate": 0.002975896846139141,
+ "loss": 6.288534164428711,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.017861567437648773,
+ "learning_rate": 0.0029756570761911147,
+ "loss": 6.258049964904785,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.019980227574706078,
+ "learning_rate": 0.0029754161294180646,
+ "loss": 6.250895023345947,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.01443127728998661,
+ "learning_rate": 0.0029751740060141177,
+ "loss": 6.212200164794922,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.014975055120885372,
+ "learning_rate": 0.0029749307061743475,
+ "loss": 6.230568885803223,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01408296450972557,
+ "learning_rate": 0.0029746862300947773,
+ "loss": 6.243950366973877,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.011649327352643013,
+ "learning_rate": 0.0029744405779723766,
+ "loss": 6.197744846343994,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.013392188586294651,
+ "learning_rate": 0.002974193750005063,
+ "loss": 6.207277297973633,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.010902997106313705,
+ "learning_rate": 0.002973945746391701,
+ "loss": 6.166373252868652,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.012915392406284809,
+ "learning_rate": 0.0029736965673321032,
+ "loss": 6.186316013336182,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.010772830806672573,
+ "learning_rate": 0.0029734462130270273,
+ "loss": 6.172578811645508,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.010274911299347878,
+ "learning_rate": 0.002973194683678181,
+ "loss": 6.206851005554199,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.013210494071245193,
+ "learning_rate": 0.0029729419794882164,
+ "loss": 6.21115779876709,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.013372313231229782,
+ "learning_rate": 0.0029726881006607326,
+ "loss": 6.18442440032959,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.008105908520519733,
+ "learning_rate": 0.0029724330474002743,
+ "loss": 6.204354286193848,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.014167108573019505,
+ "learning_rate": 0.0029721768199123346,
+ "loss": 6.180662155151367,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.008339803665876389,
+ "learning_rate": 0.002971919418403351,
+ "loss": 6.150079727172852,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.015939386561512947,
+ "learning_rate": 0.0029716608430807073,
+ "loss": 6.175203323364258,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.010777109302580357,
+ "learning_rate": 0.0029714010941527327,
+ "loss": 6.166106224060059,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.011274735443294048,
+ "learning_rate": 0.002971140171828702,
+ "loss": 6.163111686706543,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013626808300614357,
+ "learning_rate": 0.002970878076318836,
+ "loss": 6.140850067138672,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.016907762736082077,
+ "learning_rate": 0.0029706148078343,
+ "loss": 6.192012786865234,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.018817441537976265,
+ "learning_rate": 0.002970350366587205,
+ "loss": 6.162238121032715,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.02829768694937229,
+ "learning_rate": 0.0029700847527906068,
+ "loss": 6.159066200256348,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.02034052088856697,
+ "learning_rate": 0.002969817966658505,
+ "loss": 6.164158344268799,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.01816369965672493,
+ "learning_rate": 0.0029695500084058447,
+ "loss": 6.132316589355469,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.013685261830687523,
+ "learning_rate": 0.0029692808782485146,
+ "loss": 6.152181625366211,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.009791149757802486,
+ "learning_rate": 0.0029690105764033485,
+ "loss": 6.138544082641602,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.014827069826424122,
+ "learning_rate": 0.002968739103088124,
+ "loss": 6.135766983032227,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.01391249243170023,
+ "learning_rate": 0.0029684664585215618,
+ "loss": 6.146172523498535,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.012729899026453495,
+ "learning_rate": 0.002968192642923326,
+ "loss": 6.122613430023193,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.014175458811223507,
+ "learning_rate": 0.002967917656514026,
+ "loss": 6.12574577331543,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.010958840139210224,
+ "learning_rate": 0.002967641499515213,
+ "loss": 6.111088752746582,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.015380816534161568,
+ "learning_rate": 0.0029673641721493817,
+ "loss": 6.100615501403809,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.025339096784591675,
+ "learning_rate": 0.00296708567463997,
+ "loss": 6.139659881591797,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.03695995733141899,
+ "learning_rate": 0.002966806007211358,
+ "loss": 6.160996437072754,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.0408046580851078,
+ "learning_rate": 0.0029665251700888686,
+ "loss": 6.173980712890625,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.025022024288773537,
+ "learning_rate": 0.0029662431634987674,
+ "loss": 6.145794868469238,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.027254154905676842,
+ "learning_rate": 0.002965959987668262,
+ "loss": 6.16928243637085,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.06191903352737427,
+ "learning_rate": 0.0029656756428255017,
+ "loss": 6.138525009155273,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0348956398665905,
+ "learning_rate": 0.002965390129199579,
+ "loss": 6.185477256774902,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.020859677344560623,
+ "learning_rate": 0.002965103447020526,
+ "loss": 6.148846626281738,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.022810814902186394,
+ "learning_rate": 0.0029648155965193174,
+ "loss": 6.1701178550720215,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.017981817945837975,
+ "learning_rate": 0.00296452657792787,
+ "loss": 6.1323771476745605,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.014905866235494614,
+ "learning_rate": 0.00296423639147904,
+ "loss": 6.123426914215088,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.01463468000292778,
+ "learning_rate": 0.002963945037406626,
+ "loss": 6.146122932434082,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.021616583690047264,
+ "learning_rate": 0.002963652515945366,
+ "loss": 6.093459606170654,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.01550968922674656,
+ "learning_rate": 0.00296335882733094,
+ "loss": 6.114929676055908,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.014597428031265736,
+ "learning_rate": 0.0029630639717999675,
+ "loss": 6.113945960998535,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.018249280750751495,
+ "learning_rate": 0.002962767949590008,
+ "loss": 6.130507469177246,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.014160148799419403,
+ "learning_rate": 0.0029624707609395614,
+ "loss": 6.086946487426758,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.015563832595944405,
+ "learning_rate": 0.002962172406088067,
+ "loss": 6.083332538604736,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.012240015901625156,
+ "learning_rate": 0.002961872885275904,
+ "loss": 6.0822906494140625,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.0141651201993227,
+ "learning_rate": 0.002961572198744391,
+ "loss": 6.049562454223633,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.010004645213484764,
+ "learning_rate": 0.0029612703467357866,
+ "loss": 6.067063808441162,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.011790835298597813,
+ "learning_rate": 0.0029609673294932856,
+ "loss": 6.062193870544434,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.011088060215115547,
+ "learning_rate": 0.002960663147261025,
+ "loss": 6.070950031280518,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.010387620888650417,
+ "learning_rate": 0.002960357800284078,
+ "loss": 6.032423973083496,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.012074458412826061,
+ "learning_rate": 0.002960051288808458,
+ "loss": 6.066788673400879,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.097111701965332,
+ "eval_runtime": 51.8742,
+ "eval_samples_per_second": 47.075,
+ "eval_steps_per_second": 1.484,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.013429952785372734,
+ "learning_rate": 0.0029597436130811153,
+ "loss": 6.022172451019287,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.01764705777168274,
+ "learning_rate": 0.0029594347733499387,
+ "loss": 6.029745101928711,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.02403099089860916,
+ "learning_rate": 0.002959124769863755,
+ "loss": 6.04707145690918,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.029030559584498405,
+ "learning_rate": 0.002958813602872328,
+ "loss": 6.05479097366333,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.029980644583702087,
+ "learning_rate": 0.0029585012726263595,
+ "loss": 6.068657875061035,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.021424276754260063,
+ "learning_rate": 0.0029581877793774885,
+ "loss": 6.062008857727051,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.018076395615935326,
+ "learning_rate": 0.002957873123378291,
+ "loss": 6.023279190063477,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.017688792198896408,
+ "learning_rate": 0.002957557304882279,
+ "loss": 6.014641284942627,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.02313626930117607,
+ "learning_rate": 0.0029572403241439026,
+ "loss": 6.026420593261719,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.04045062139630318,
+ "learning_rate": 0.0029569221814185467,
+ "loss": 6.007199764251709,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.015792209655046463,
+ "learning_rate": 0.0029566028769625334,
+ "loss": 6.017209053039551,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.01990479789674282,
+ "learning_rate": 0.002956282411033121,
+ "loss": 5.982841491699219,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.01191043108701706,
+ "learning_rate": 0.002955960783888503,
+ "loss": 6.0035929679870605,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.01732831820845604,
+ "learning_rate": 0.002955637995787807,
+ "loss": 6.025006294250488,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.032948266714811325,
+ "learning_rate": 0.0029553140469910995,
+ "loss": 6.007736682891846,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.011192727833986282,
+ "learning_rate": 0.00295498893775938,
+ "loss": 6.0022687911987305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014646840281784534,
+ "learning_rate": 0.0029546626683545824,
+ "loss": 5.976479530334473,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01589796133339405,
+ "learning_rate": 0.002954335239039576,
+ "loss": 5.980949878692627,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.015129734762012959,
+ "learning_rate": 0.0029540066500781657,
+ "loss": 6.0073394775390625,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.011140518821775913,
+ "learning_rate": 0.002953676901735088,
+ "loss": 5.976587295532227,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.0109480582177639,
+ "learning_rate": 0.0029533459942760174,
+ "loss": 6.001427173614502,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.013346790336072445,
+ "learning_rate": 0.0029530139279675583,
+ "loss": 5.96919059753418,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.012474581599235535,
+ "learning_rate": 0.002952680703077251,
+ "loss": 5.970426559448242,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01140566635876894,
+ "learning_rate": 0.002952346319873569,
+ "loss": 5.989171504974365,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.0137029392644763,
+ "learning_rate": 0.0029520107786259187,
+ "loss": 6.010115623474121,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015517761930823326,
+ "learning_rate": 0.002951674079604639,
+ "loss": 6.003649711608887,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.02173629403114319,
+ "learning_rate": 0.0029513362230810035,
+ "loss": 5.967074871063232,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02683914266526699,
+ "learning_rate": 0.002950997209327216,
+ "loss": 5.994434833526611,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.025865858420729637,
+ "learning_rate": 0.002950657038616415,
+ "loss": 6.010523796081543,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02544897235929966,
+ "learning_rate": 0.002950315711222668,
+ "loss": 5.994078159332275,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.028426019474864006,
+ "learning_rate": 0.0029499732274209777,
+ "loss": 6.020925521850586,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.026174750179052353,
+ "learning_rate": 0.0029496295874872768,
+ "loss": 5.985616683959961,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.02802138403058052,
+ "learning_rate": 0.00294928479169843,
+ "loss": 5.973659038543701,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.019348938018083572,
+ "learning_rate": 0.002948938840332233,
+ "loss": 5.990137577056885,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.016294041648507118,
+ "learning_rate": 0.0029485917336674115,
+ "loss": 5.962082862854004,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.016167785972356796,
+ "learning_rate": 0.002948243471983625,
+ "loss": 5.9594573974609375,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.014119232073426247,
+ "learning_rate": 0.00294789405556146,
+ "loss": 5.894467353820801,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.013805347494781017,
+ "learning_rate": 0.002947543484682436,
+ "loss": 5.964014530181885,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.014496497809886932,
+ "learning_rate": 0.002947191759629001,
+ "loss": 5.989920616149902,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.014813800342381,
+ "learning_rate": 0.0029468388806845333,
+ "loss": 5.9669413566589355,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.01153622381389141,
+ "learning_rate": 0.0029464848481333423,
+ "loss": 5.918018341064453,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.010525977239012718,
+ "learning_rate": 0.0029461296622606647,
+ "loss": 5.892477989196777,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.011410463601350784,
+ "learning_rate": 0.002945773323352667,
+ "loss": 5.925249099731445,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.010949424467980862,
+ "learning_rate": 0.0029454158316964464,
+ "loss": 5.90645694732666,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.00962804164737463,
+ "learning_rate": 0.002945057187580026,
+ "loss": 5.917054176330566,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.011517546139657497,
+ "learning_rate": 0.0029446973912923597,
+ "loss": 5.919445514678955,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01567666605114937,
+ "learning_rate": 0.0029443364431233285,
+ "loss": 5.884111404418945,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.016781140118837357,
+ "learning_rate": 0.002943974343363742,
+ "loss": 5.9252119064331055,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.02244546078145504,
+ "learning_rate": 0.0029436110923053375,
+ "loss": 5.933143615722656,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.028186792507767677,
+ "learning_rate": 0.00294324669024078,
+ "loss": 5.9079132080078125,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.028585223481059074,
+ "learning_rate": 0.002942881137463661,
+ "loss": 5.925262451171875,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.02781485952436924,
+ "learning_rate": 0.0029425144342685003,
+ "loss": 5.963315010070801,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.021486928686499596,
+ "learning_rate": 0.002942146580950744,
+ "loss": 5.930115699768066,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.016890818253159523,
+ "learning_rate": 0.002941777577806765,
+ "loss": 5.921439170837402,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.0169126745313406,
+ "learning_rate": 0.0029414074251338618,
+ "loss": 5.928220272064209,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.014659176580607891,
+ "learning_rate": 0.002941036123230261,
+ "loss": 5.934546947479248,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.015337883494794369,
+ "learning_rate": 0.002940663672395112,
+ "loss": 5.890145301818848,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01254988368600607,
+ "learning_rate": 0.002940290072928493,
+ "loss": 5.875492095947266,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.014658362604677677,
+ "learning_rate": 0.0029399153251314066,
+ "loss": 5.901272773742676,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.017342250794172287,
+ "learning_rate": 0.002939539429305779,
+ "loss": 5.875853538513184,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.016825608909130096,
+ "learning_rate": 0.0029391623857544643,
+ "loss": 5.941972732543945,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.016716884449124336,
+ "learning_rate": 0.002938784194781238,
+ "loss": 5.912367343902588,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.021707067266106606,
+ "learning_rate": 0.002938404856690803,
+ "loss": 5.895319938659668,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.030519135296344757,
+ "learning_rate": 0.002938024371788785,
+ "loss": 5.897165298461914,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.02572188340127468,
+ "learning_rate": 0.002937642740381733,
+ "loss": 5.879164695739746,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.016286717727780342,
+ "learning_rate": 0.002937259962777122,
+ "loss": 5.841421127319336,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.020656082779169083,
+ "learning_rate": 0.002936876039283347,
+ "loss": 5.9222259521484375,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.015350459143519402,
+ "learning_rate": 0.0029364909702097298,
+ "loss": 5.885847091674805,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01708570495247841,
+ "learning_rate": 0.0029361047558665136,
+ "loss": 5.872044563293457,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.013537376187741756,
+ "learning_rate": 0.0029357173965648634,
+ "loss": 5.866434097290039,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.013208637945353985,
+ "learning_rate": 0.002935328892616868,
+ "loss": 5.8624491691589355,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.012249554507434368,
+ "learning_rate": 0.002934939244335538,
+ "loss": 5.833232879638672,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.90985107421875,
+ "eval_runtime": 51.0696,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.01382362935692072,
+ "learning_rate": 0.0029345484520348062,
+ "loss": 5.835695266723633,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.012947693467140198,
+ "learning_rate": 0.0029341565160295262,
+ "loss": 5.826190948486328,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.013653872534632683,
+ "learning_rate": 0.002933763436635475,
+ "loss": 5.831789016723633,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.027596980333328247,
+ "learning_rate": 0.0029333692141693484,
+ "loss": 5.839963436126709,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01637990213930607,
+ "learning_rate": 0.0029329738489487648,
+ "loss": 5.805248260498047,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.02169182151556015,
+ "learning_rate": 0.0029325773412922627,
+ "loss": 5.838248252868652,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.03518201410770416,
+ "learning_rate": 0.0029321796915193015,
+ "loss": 5.834336280822754,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04983877018094063,
+ "learning_rate": 0.0029317808999502603,
+ "loss": 5.879856109619141,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05085664615035057,
+ "learning_rate": 0.0029313809669064377,
+ "loss": 5.946531295776367,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.03314858302474022,
+ "learning_rate": 0.0029309798927100532,
+ "loss": 5.909746170043945,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.034118980169296265,
+ "learning_rate": 0.0029305776776842453,
+ "loss": 5.903947830200195,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.03401247411966324,
+ "learning_rate": 0.0029301743221530705,
+ "loss": 5.936162948608398,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.043827857822179794,
+ "learning_rate": 0.0029297698264415047,
+ "loss": 5.953949451446533,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.06281330436468124,
+ "learning_rate": 0.002929364190875444,
+ "loss": 5.988225936889648,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.07216024398803711,
+ "learning_rate": 0.0029289574157817014,
+ "loss": 6.187741279602051,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.055205658078193665,
+ "learning_rate": 0.002928549501488008,
+ "loss": 6.177831649780273,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.04834286496043205,
+ "learning_rate": 0.0029281404483230133,
+ "loss": 6.209604263305664,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.0447777584195137,
+ "learning_rate": 0.0029277302566162837,
+ "loss": 6.146507263183594,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.05729430541396141,
+ "learning_rate": 0.0029273189266983035,
+ "loss": 6.280686378479004,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.0410807766020298,
+ "learning_rate": 0.002926906458900473,
+ "loss": 6.246824264526367,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.044607777148485184,
+ "learning_rate": 0.0029264928535551104,
+ "loss": 6.19956111907959,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.05141043663024902,
+ "learning_rate": 0.0029260781109954515,
+ "loss": 6.118578910827637,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.029248764738440514,
+ "learning_rate": 0.0029256622315556448,
+ "loss": 6.155360221862793,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.028403814882040024,
+ "learning_rate": 0.0029252452155707582,
+ "loss": 6.097352027893066,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.025872666388750076,
+ "learning_rate": 0.002924827063376773,
+ "loss": 6.072485446929932,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.018447555601596832,
+ "learning_rate": 0.0029244077753105874,
+ "loss": 6.077045440673828,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.023166058585047722,
+ "learning_rate": 0.0029239873517100147,
+ "loss": 6.040050983428955,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.019579261541366577,
+ "learning_rate": 0.002923565792913782,
+ "loss": 6.039424896240234,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.018369073048233986,
+ "learning_rate": 0.002923143099261531,
+ "loss": 6.021180152893066,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.014188461937010288,
+ "learning_rate": 0.0029227192710938195,
+ "loss": 5.9668049812316895,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.019567398354411125,
+ "learning_rate": 0.0029222943087521182,
+ "loss": 5.989189147949219,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.014190621674060822,
+ "learning_rate": 0.0029218682125788102,
+ "loss": 5.942726135253906,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01375526562333107,
+ "learning_rate": 0.002921440982917195,
+ "loss": 5.942777633666992,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.014203300699591637,
+ "learning_rate": 0.0029210126201114836,
+ "loss": 5.927928924560547,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.014231069944798946,
+ "learning_rate": 0.0029205831245067994,
+ "loss": 5.923510551452637,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.012771515175700188,
+ "learning_rate": 0.0029201524964491796,
+ "loss": 5.908803939819336,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.013399476185441017,
+ "learning_rate": 0.0029197207362855735,
+ "loss": 5.890079021453857,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.025862423703074455,
+ "learning_rate": 0.0029192878443638427,
+ "loss": 5.919907569885254,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.011315986514091492,
+ "learning_rate": 0.0029188538210327595,
+ "loss": 5.901529788970947,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.01235947571694851,
+ "learning_rate": 0.00291841866664201,
+ "loss": 5.903595924377441,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.012175127863883972,
+ "learning_rate": 0.0029179823815421893,
+ "loss": 5.868306636810303,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01195445004850626,
+ "learning_rate": 0.0029175449660848052,
+ "loss": 5.899898529052734,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.012801481410861015,
+ "learning_rate": 0.0029171064206222745,
+ "loss": 5.890621662139893,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.011909718625247478,
+ "learning_rate": 0.002916666745507926,
+ "loss": 5.86421012878418,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01162486057728529,
+ "learning_rate": 0.0029162259410959977,
+ "loss": 5.869139671325684,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.010923407040536404,
+ "learning_rate": 0.0029157840077416374,
+ "loss": 5.869789123535156,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.011943561024963856,
+ "learning_rate": 0.002915340945800904,
+ "loss": 5.873437881469727,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.010295520536601543,
+ "learning_rate": 0.002914896755630763,
+ "loss": 5.819953918457031,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.0093373479321599,
+ "learning_rate": 0.002914451437589092,
+ "loss": 5.829481601715088,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.01182442344725132,
+ "learning_rate": 0.002914004992034674,
+ "loss": 5.855249404907227,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.00939213763922453,
+ "learning_rate": 0.0029135574193272034,
+ "loss": 5.83876371383667,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.011648556217551231,
+ "learning_rate": 0.0029131087198272813,
+ "loss": 5.824906826019287,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.011227169074118137,
+ "learning_rate": 0.002912658893896416,
+ "loss": 5.847869873046875,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.011084851808845997,
+ "learning_rate": 0.002912207941897025,
+ "loss": 5.839183807373047,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.009923825040459633,
+ "learning_rate": 0.002911755864192432,
+ "loss": 5.818821907043457,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.011300318874418736,
+ "learning_rate": 0.002911302661146867,
+ "loss": 5.819517135620117,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.009915645234286785,
+ "learning_rate": 0.0029108483331254684,
+ "loss": 5.776939392089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.008965205401182175,
+ "learning_rate": 0.00291039288049428,
+ "loss": 5.824148178100586,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.011025927029550076,
+ "learning_rate": 0.0029099363036202515,
+ "loss": 5.832704544067383,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.010143411345779896,
+ "learning_rate": 0.0029094786028712387,
+ "loss": 5.802353382110596,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.010021256282925606,
+ "learning_rate": 0.002909019778616003,
+ "loss": 5.802136421203613,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.011356706731021404,
+ "learning_rate": 0.002908559831224211,
+ "loss": 5.807962417602539,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.010525298304855824,
+ "learning_rate": 0.002908098761066433,
+ "loss": 5.787661552429199,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.011147079057991505,
+ "learning_rate": 0.0029076365685141463,
+ "loss": 5.801015377044678,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.008777810260653496,
+ "learning_rate": 0.00290717325393973,
+ "loss": 5.779858589172363,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.011423609219491482,
+ "learning_rate": 0.002906708817716469,
+ "loss": 5.777166843414307,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.008639856241643429,
+ "learning_rate": 0.00290624326021855,
+ "loss": 5.747189521789551,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.009628898464143276,
+ "learning_rate": 0.002905776581821066,
+ "loss": 5.7666916847229,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.011988377198576927,
+ "learning_rate": 0.0029053087829000097,
+ "loss": 5.777951240539551,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.01693224348127842,
+ "learning_rate": 0.00290483986383228,
+ "loss": 5.818393707275391,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.020264796912670135,
+ "learning_rate": 0.002904369824995674,
+ "loss": 5.747611045837402,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.018283747136592865,
+ "learning_rate": 0.0029038986667688964,
+ "loss": 5.789663314819336,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.825676918029785,
+ "eval_runtime": 57.6908,
+ "eval_samples_per_second": 42.329,
+ "eval_steps_per_second": 1.335,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.011447477154433727,
+ "learning_rate": 0.002903426389531549,
+ "loss": 5.724975109100342,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.012539781630039215,
+ "learning_rate": 0.002902952993664138,
+ "loss": 5.747196674346924,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.012179837562143803,
+ "learning_rate": 0.002902478479548069,
+ "loss": 5.748995780944824,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.019896456971764565,
+ "learning_rate": 0.00290200284756565,
+ "loss": 5.751996994018555,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.022261742502450943,
+ "learning_rate": 0.002901526098100089,
+ "loss": 5.7516984939575195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02683536894619465,
+ "learning_rate": 0.0029010482315354946,
+ "loss": 5.716037750244141,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.018057072535157204,
+ "learning_rate": 0.0029005692482568747,
+ "loss": 5.778608322143555,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.011646891944110394,
+ "learning_rate": 0.002900089148650138,
+ "loss": 5.7308268547058105,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02230495773255825,
+ "learning_rate": 0.002899607933102092,
+ "loss": 5.768552780151367,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.01212464738637209,
+ "learning_rate": 0.002899125602000443,
+ "loss": 5.733015060424805,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.011764907278120518,
+ "learning_rate": 0.0028986421557337973,
+ "loss": 5.748104095458984,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.01593147963285446,
+ "learning_rate": 0.002898157594691658,
+ "loss": 5.701960563659668,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.016333773732185364,
+ "learning_rate": 0.0028976719192644277,
+ "loss": 5.716808795928955,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.017547868192195892,
+ "learning_rate": 0.002897185129843406,
+ "loss": 5.724377632141113,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.01662801019847393,
+ "learning_rate": 0.002896697226820791,
+ "loss": 5.737432956695557,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.015781691297888756,
+ "learning_rate": 0.002896208210589676,
+ "loss": 5.7376203536987305,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.014363214373588562,
+ "learning_rate": 0.0028957180815440534,
+ "loss": 5.7342610359191895,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.015048054046928883,
+ "learning_rate": 0.0028952268400788115,
+ "loss": 5.715073585510254,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.013753226026892662,
+ "learning_rate": 0.002894734486589734,
+ "loss": 5.73264217376709,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.015562563203275204,
+ "learning_rate": 0.002894241021473502,
+ "loss": 5.7072038650512695,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.01520819216966629,
+ "learning_rate": 0.0028937464451276916,
+ "loss": 5.683300018310547,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.013024724088609219,
+ "learning_rate": 0.002893250757950773,
+ "loss": 5.726273536682129,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.012760856188833714,
+ "learning_rate": 0.0028927539603421128,
+ "loss": 5.7249064445495605,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.013542579486966133,
+ "learning_rate": 0.0028922560527019723,
+ "loss": 5.70849084854126,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.014801754616200924,
+ "learning_rate": 0.002891757035431506,
+ "loss": 5.725185394287109,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.013000023551285267,
+ "learning_rate": 0.0028912569089327637,
+ "loss": 5.705799102783203,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.012919292785227299,
+ "learning_rate": 0.002890755673608688,
+ "loss": 5.674324989318848,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.015904201194643974,
+ "learning_rate": 0.002890253329863115,
+ "loss": 5.680788993835449,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.015467721968889236,
+ "learning_rate": 0.0028897498781007744,
+ "loss": 5.707531929016113,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.015157260000705719,
+ "learning_rate": 0.002889245318727288,
+ "loss": 5.689894676208496,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.014204306527972221,
+ "learning_rate": 0.00288873965214917,
+ "loss": 5.713001251220703,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.022045845165848732,
+ "learning_rate": 0.0028882328787738273,
+ "loss": 5.657631874084473,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.013380175456404686,
+ "learning_rate": 0.0028877249990095573,
+ "loss": 5.702192306518555,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.016698408871889114,
+ "learning_rate": 0.0028872160132655503,
+ "loss": 5.695206642150879,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.020001472905278206,
+ "learning_rate": 0.0028867059219518866,
+ "loss": 5.731742858886719,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.027350103482604027,
+ "learning_rate": 0.0028861947254795375,
+ "loss": 5.712314605712891,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.021862946450710297,
+ "learning_rate": 0.002885682424260365,
+ "loss": 5.697043418884277,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.019216977059841156,
+ "learning_rate": 0.0028851690187071204,
+ "loss": 5.698813438415527,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.02131095714867115,
+ "learning_rate": 0.002884654509233446,
+ "loss": 5.6918864250183105,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.016583887860178947,
+ "learning_rate": 0.0028841388962538724,
+ "loss": 5.711456298828125,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.016998684033751488,
+ "learning_rate": 0.0028836221801838194,
+ "loss": 5.675747871398926,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.01705755852162838,
+ "learning_rate": 0.0028831043614395966,
+ "loss": 5.697177410125732,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.014604352414608002,
+ "learning_rate": 0.0028825854404384,
+ "loss": 5.681950092315674,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.013561096973717213,
+ "learning_rate": 0.0028820654175983162,
+ "loss": 5.67357063293457,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.021149173378944397,
+ "learning_rate": 0.0028815442933383174,
+ "loss": 5.709477424621582,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.01431041955947876,
+ "learning_rate": 0.0028810220680782644,
+ "loss": 5.7012200355529785,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.01429709978401661,
+ "learning_rate": 0.0028804987422389045,
+ "loss": 5.689794540405273,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.032025713473558426,
+ "learning_rate": 0.0028799743162418712,
+ "loss": 5.6713762283325195,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.013168714940547943,
+ "learning_rate": 0.002879448790509686,
+ "loss": 5.685478210449219,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.014088666997849941,
+ "learning_rate": 0.0028789221654657556,
+ "loss": 5.645514965057373,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.013320053927600384,
+ "learning_rate": 0.002878394441534372,
+ "loss": 5.678678512573242,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.022180087864398956,
+ "learning_rate": 0.002877865619140712,
+ "loss": 5.660525321960449,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.01271889265626669,
+ "learning_rate": 0.002877335698710839,
+ "loss": 5.652164459228516,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.011617431417107582,
+ "learning_rate": 0.002876804680671701,
+ "loss": 5.6389946937561035,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.012040726840496063,
+ "learning_rate": 0.0028762725654511282,
+ "loss": 5.628744602203369,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.013148069381713867,
+ "learning_rate": 0.0028757393534778377,
+ "loss": 5.69610595703125,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.01826276071369648,
+ "learning_rate": 0.0028752050451814276,
+ "loss": 5.634238243103027,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.020393460988998413,
+ "learning_rate": 0.0028746696409923805,
+ "loss": 5.656968593597412,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.017072148621082306,
+ "learning_rate": 0.0028741331413420627,
+ "loss": 5.668219089508057,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.01694098301231861,
+ "learning_rate": 0.0028735955466627217,
+ "loss": 5.670073509216309,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.022760670632123947,
+ "learning_rate": 0.002873056857387488,
+ "loss": 5.649288177490234,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.03613442927598953,
+ "learning_rate": 0.0028725170739503734,
+ "loss": 5.723211288452148,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.04232329502701759,
+ "learning_rate": 0.002871976196786272,
+ "loss": 5.757920742034912,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.044210851192474365,
+ "learning_rate": 0.002871434226330959,
+ "loss": 5.702426910400391,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.044886402785778046,
+ "learning_rate": 0.0028708911630210905,
+ "loss": 5.742881774902344,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.022805195301771164,
+ "learning_rate": 0.002870347007294202,
+ "loss": 5.716353416442871,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02219395525753498,
+ "learning_rate": 0.00286980175958871,
+ "loss": 5.68503475189209,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.019240057095885277,
+ "learning_rate": 0.002869255420343911,
+ "loss": 5.700656890869141,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.01957295648753643,
+ "learning_rate": 0.0028687079899999807,
+ "loss": 5.70595645904541,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016326555982232094,
+ "learning_rate": 0.0028681594689979736,
+ "loss": 5.697229385375977,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015815885737538338,
+ "learning_rate": 0.0028676098577798234,
+ "loss": 5.7020792961120605,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.015865175053477287,
+ "learning_rate": 0.002867059156788342,
+ "loss": 5.648272514343262,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.7417449951171875,
+ "eval_runtime": 51.2263,
+ "eval_samples_per_second": 47.671,
+ "eval_steps_per_second": 1.503,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.016063183546066284,
+ "learning_rate": 0.0028665073664672185,
+ "loss": 5.65411376953125,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.014323881827294827,
+ "learning_rate": 0.0028659544872610214,
+ "loss": 5.629649639129639,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.014005238190293312,
+ "learning_rate": 0.002865400519615195,
+ "loss": 5.655135631561279,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.01385372132062912,
+ "learning_rate": 0.0028648454639760615,
+ "loss": 5.632213115692139,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.014073756523430347,
+ "learning_rate": 0.0028642893207908187,
+ "loss": 5.626431465148926,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.019551943987607956,
+ "learning_rate": 0.0028637320905075406,
+ "loss": 5.595447063446045,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.017571093514561653,
+ "learning_rate": 0.002863173773575179,
+ "loss": 5.661581039428711,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.01691303588449955,
+ "learning_rate": 0.0028626143704435585,
+ "loss": 5.641129493713379,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.01584242656826973,
+ "learning_rate": 0.0028620538815633807,
+ "loss": 5.640626907348633,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.014592235907912254,
+ "learning_rate": 0.0028614923073862213,
+ "loss": 5.611204147338867,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.016636405140161514,
+ "learning_rate": 0.0028609296483645304,
+ "loss": 5.614716053009033,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.01370833721011877,
+ "learning_rate": 0.002860365904951632,
+ "loss": 5.633011817932129,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.013218311592936516,
+ "learning_rate": 0.002859801077601725,
+ "loss": 5.588958740234375,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.016971616074442863,
+ "learning_rate": 0.0028592351667698786,
+ "loss": 5.635064125061035,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.015524905174970627,
+ "learning_rate": 0.0028586681729120393,
+ "loss": 5.611454963684082,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.014769157394766808,
+ "learning_rate": 0.0028581000964850215,
+ "loss": 5.62175178527832,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.014313343912363052,
+ "learning_rate": 0.002857530937946516,
+ "loss": 5.603628158569336,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.016748322173953056,
+ "learning_rate": 0.002856960697755083,
+ "loss": 5.617765426635742,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.013503300957381725,
+ "learning_rate": 0.0028563893763701538,
+ "loss": 5.636135578155518,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.012275655753910542,
+ "learning_rate": 0.002855816974252033,
+ "loss": 5.552533149719238,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.01288773212581873,
+ "learning_rate": 0.002855243491861894,
+ "loss": 5.620527267456055,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.013687008060514927,
+ "learning_rate": 0.002854668929661781,
+ "loss": 5.618595123291016,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.015678439289331436,
+ "learning_rate": 0.0028540932881146092,
+ "loss": 5.604031562805176,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.013567988760769367,
+ "learning_rate": 0.002853516567684162,
+ "loss": 5.549149513244629,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0157297495752573,
+ "learning_rate": 0.002852938768835092,
+ "loss": 5.618505477905273,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.011867019347846508,
+ "learning_rate": 0.0028523598920329233,
+ "loss": 5.579049110412598,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.01398563478142023,
+ "learning_rate": 0.0028517799377440443,
+ "loss": 5.597678184509277,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.015525273978710175,
+ "learning_rate": 0.0028511989064357148,
+ "loss": 5.601663589477539,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.01588466763496399,
+ "learning_rate": 0.002850616798576061,
+ "loss": 5.5650434494018555,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.014276969246566296,
+ "learning_rate": 0.0028500336146340772,
+ "loss": 5.607125282287598,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.013253689743578434,
+ "learning_rate": 0.002849449355079623,
+ "loss": 5.6299824714660645,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.01176661066710949,
+ "learning_rate": 0.0028488640203834263,
+ "loss": 5.560839653015137,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.011712565086781979,
+ "learning_rate": 0.0028482776110170813,
+ "loss": 5.618312835693359,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.013515647500753403,
+ "learning_rate": 0.002847690127453047,
+ "loss": 5.566764831542969,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.020177524536848068,
+ "learning_rate": 0.0028471015701646485,
+ "loss": 5.554571628570557,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.029197178781032562,
+ "learning_rate": 0.0028465119396260757,
+ "loss": 5.6097259521484375,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.027947140857577324,
+ "learning_rate": 0.0028459212363123824,
+ "loss": 5.65598201751709,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.021089809015393257,
+ "learning_rate": 0.0028453294606994887,
+ "loss": 5.67195987701416,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.019936271011829376,
+ "learning_rate": 0.0028447366132641773,
+ "loss": 5.609482765197754,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.0176444910466671,
+ "learning_rate": 0.002844142694484094,
+ "loss": 5.60432767868042,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.023154428228735924,
+ "learning_rate": 0.0028435477048377495,
+ "loss": 5.616270065307617,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.018391480669379234,
+ "learning_rate": 0.002842951644804515,
+ "loss": 5.600588321685791,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.013716557994484901,
+ "learning_rate": 0.002842354514864627,
+ "loss": 5.595095634460449,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.014372860081493855,
+ "learning_rate": 0.002841756315499181,
+ "loss": 5.584517478942871,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.014930845238268375,
+ "learning_rate": 0.002841157047190135,
+ "loss": 5.587005138397217,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.012631502002477646,
+ "learning_rate": 0.0028405567104203095,
+ "loss": 5.594094276428223,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.014677002094686031,
+ "learning_rate": 0.0028399553056733854,
+ "loss": 5.569582939147949,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.015558961778879166,
+ "learning_rate": 0.0028393528334339025,
+ "loss": 5.596535682678223,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.014322095550596714,
+ "learning_rate": 0.0028387492941872624,
+ "loss": 5.570324420928955,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.012913414277136326,
+ "learning_rate": 0.002838144688419726,
+ "loss": 5.573002338409424,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.01251310296356678,
+ "learning_rate": 0.0028375390166184137,
+ "loss": 5.550327777862549,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.013847017660737038,
+ "learning_rate": 0.0028369322792713026,
+ "loss": 5.54567813873291,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.013789358548820019,
+ "learning_rate": 0.0028363244768672313,
+ "loss": 5.571495056152344,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.01236631814390421,
+ "learning_rate": 0.0028357156098958954,
+ "loss": 5.569730758666992,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.013210169039666653,
+ "learning_rate": 0.002835105678847847,
+ "loss": 5.53984260559082,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.01604137383401394,
+ "learning_rate": 0.0028344946842144975,
+ "loss": 5.530598163604736,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.01708815060555935,
+ "learning_rate": 0.002833882626488114,
+ "loss": 5.529256820678711,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.017222166061401367,
+ "learning_rate": 0.0028332695061618197,
+ "loss": 5.564949035644531,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.01803218014538288,
+ "learning_rate": 0.002832655323729595,
+ "loss": 5.560049057006836,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.01658225618302822,
+ "learning_rate": 0.002832040079686276,
+ "loss": 5.553183555603027,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.016834143549203873,
+ "learning_rate": 0.002831423774527554,
+ "loss": 5.573678970336914,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.01856408454477787,
+ "learning_rate": 0.002830806408749974,
+ "loss": 5.539497375488281,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.026909375563263893,
+ "learning_rate": 0.002830187982850936,
+ "loss": 5.551433563232422,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.016046348959207535,
+ "learning_rate": 0.002829568497328696,
+ "loss": 5.566728591918945,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.018582617864012718,
+ "learning_rate": 0.0028289479526823614,
+ "loss": 5.543226718902588,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.0211537666618824,
+ "learning_rate": 0.0028283263494118943,
+ "loss": 5.560711860656738,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.013894101604819298,
+ "learning_rate": 0.0028277036880181095,
+ "loss": 5.528890132904053,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.017581885680556297,
+ "learning_rate": 0.0028270799690026725,
+ "loss": 5.57473087310791,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.01622696965932846,
+ "learning_rate": 0.002826455192868104,
+ "loss": 5.578961372375488,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.01966337114572525,
+ "learning_rate": 0.0028258293601177743,
+ "loss": 5.513942718505859,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021581366658210754,
+ "learning_rate": 0.002825202471255906,
+ "loss": 5.573669910430908,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.020468758419156075,
+ "learning_rate": 0.002824574526787571,
+ "loss": 5.516103744506836,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.626922130584717,
+ "eval_runtime": 54.2404,
+ "eval_samples_per_second": 45.022,
+ "eval_steps_per_second": 1.42,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.014597049914300442,
+ "learning_rate": 0.0028239455272186936,
+ "loss": 5.506453514099121,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.020277054980397224,
+ "learning_rate": 0.002823315473056048,
+ "loss": 5.535350799560547,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.019103925675153732,
+ "learning_rate": 0.0028226843648072562,
+ "loss": 5.522078037261963,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.07158710062503815,
+ "learning_rate": 0.0028220522029807913,
+ "loss": 5.531142234802246,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01958007924258709,
+ "learning_rate": 0.0028214189880859746,
+ "loss": 5.484055042266846,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.0185184795409441,
+ "learning_rate": 0.002820784720632976,
+ "loss": 5.511264801025391,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01976991444826126,
+ "learning_rate": 0.0028201494011328133,
+ "loss": 5.51821231842041,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.02188604325056076,
+ "learning_rate": 0.002819513030097352,
+ "loss": 5.483095169067383,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.019405223429203033,
+ "learning_rate": 0.002818875608039305,
+ "loss": 5.539701461791992,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.016907615587115288,
+ "learning_rate": 0.0028182371354722306,
+ "loss": 5.518251419067383,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.02160538360476494,
+ "learning_rate": 0.0028175976129105364,
+ "loss": 5.509786605834961,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.023125721141695976,
+ "learning_rate": 0.002816957040869473,
+ "loss": 5.535561561584473,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.022233380004763603,
+ "learning_rate": 0.0028163154198651372,
+ "loss": 5.553831100463867,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.017227552831172943,
+ "learning_rate": 0.002815672750414473,
+ "loss": 5.547080993652344,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.017122473567724228,
+ "learning_rate": 0.0028150290330352665,
+ "loss": 5.493778228759766,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.020641913637518883,
+ "learning_rate": 0.0028143842682461498,
+ "loss": 5.536645889282227,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.017383966594934464,
+ "learning_rate": 0.0028137384565665967,
+ "loss": 5.506233215332031,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.017969606444239616,
+ "learning_rate": 0.002813091598516928,
+ "loss": 5.524741172790527,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.01936613954603672,
+ "learning_rate": 0.0028124436946183047,
+ "loss": 5.460989952087402,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.017492495477199554,
+ "learning_rate": 0.0028117947453927313,
+ "loss": 5.487619876861572,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.018786396831274033,
+ "learning_rate": 0.0028111447513630545,
+ "loss": 5.516829013824463,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.020149117335677147,
+ "learning_rate": 0.0028104937130529624,
+ "loss": 5.460541725158691,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.024668103083968163,
+ "learning_rate": 0.0028098416309869857,
+ "loss": 5.497687339782715,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.015888608992099762,
+ "learning_rate": 0.002809188505690494,
+ "loss": 5.494459629058838,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.019327623769640923,
+ "learning_rate": 0.0028085343376896997,
+ "loss": 5.486351490020752,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.026855193078517914,
+ "learning_rate": 0.0028078791275116532,
+ "loss": 5.509825706481934,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.026340370997786522,
+ "learning_rate": 0.002807222875684246,
+ "loss": 5.527161598205566,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.02200835384428501,
+ "learning_rate": 0.002806565582736208,
+ "loss": 5.507146835327148,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.020420534536242485,
+ "learning_rate": 0.0028059072491971085,
+ "loss": 5.5162506103515625,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02493203990161419,
+ "learning_rate": 0.0028052478755973545,
+ "loss": 5.5252556800842285,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.021293973550200462,
+ "learning_rate": 0.0028045874624681925,
+ "loss": 5.523589134216309,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.016593236476182938,
+ "learning_rate": 0.0028039260103417045,
+ "loss": 5.500586032867432,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.018857723101973534,
+ "learning_rate": 0.002803263519750811,
+ "loss": 5.52933406829834,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.015930436551570892,
+ "learning_rate": 0.0028025999912292677,
+ "loss": 5.48813533782959,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.019427936524152756,
+ "learning_rate": 0.002801935425311669,
+ "loss": 5.46573543548584,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.018123053014278412,
+ "learning_rate": 0.0028012698225334427,
+ "loss": 5.499445915222168,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.02035410702228546,
+ "learning_rate": 0.002800603183430853,
+ "loss": 5.513540267944336,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.025838350877165794,
+ "learning_rate": 0.0027999355085409996,
+ "loss": 5.565637588500977,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.029621612280607224,
+ "learning_rate": 0.002799266798401816,
+ "loss": 5.5619215965271,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.028948379680514336,
+ "learning_rate": 0.0027985970535520685,
+ "loss": 5.551331520080566,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.022135930135846138,
+ "learning_rate": 0.0027979262745313604,
+ "loss": 5.527722358703613,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.028884069994091988,
+ "learning_rate": 0.0027972544618801256,
+ "loss": 5.554599761962891,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.02499174140393734,
+ "learning_rate": 0.00279658161613963,
+ "loss": 5.543351173400879,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.021131042391061783,
+ "learning_rate": 0.0027959077378519755,
+ "loss": 5.496039390563965,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.02957489714026451,
+ "learning_rate": 0.002795232827560092,
+ "loss": 5.552300453186035,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.022832119837403297,
+ "learning_rate": 0.0027945568858077433,
+ "loss": 5.488651275634766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.02074701525270939,
+ "learning_rate": 0.002793879913139523,
+ "loss": 5.491244792938232,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.019919803366065025,
+ "learning_rate": 0.002793201910100856,
+ "loss": 5.483855247497559,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.01858169585466385,
+ "learning_rate": 0.002792522877237998,
+ "loss": 5.50787353515625,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02085857465863228,
+ "learning_rate": 0.002791842815098032,
+ "loss": 5.512872695922852,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.01407450158149004,
+ "learning_rate": 0.002791161724228872,
+ "loss": 5.472590446472168,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.01672944612801075,
+ "learning_rate": 0.002790479605179262,
+ "loss": 5.469183921813965,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.016565700992941856,
+ "learning_rate": 0.0027897964584987713,
+ "loss": 5.493677139282227,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.017017360776662827,
+ "learning_rate": 0.0027891122847377996,
+ "loss": 5.464931488037109,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.019842486828565598,
+ "learning_rate": 0.002788427084447574,
+ "loss": 5.461293697357178,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.016210904344916344,
+ "learning_rate": 0.0027877408581801467,
+ "loss": 5.454716682434082,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.02064039371907711,
+ "learning_rate": 0.002787053606488398,
+ "loss": 5.466576099395752,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.018198711797595024,
+ "learning_rate": 0.0027863653299260356,
+ "loss": 5.447747230529785,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.015937626361846924,
+ "learning_rate": 0.00278567602904759,
+ "loss": 5.5055036544799805,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.01687259040772915,
+ "learning_rate": 0.002784985704408419,
+ "loss": 5.422250747680664,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.02622046135365963,
+ "learning_rate": 0.002784294356564705,
+ "loss": 5.466894626617432,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.018672186881303787,
+ "learning_rate": 0.0027836019860734535,
+ "loss": 5.48300838470459,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.02166241593658924,
+ "learning_rate": 0.0027829085934924958,
+ "loss": 5.477658271789551,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018426023423671722,
+ "learning_rate": 0.0027822141793804856,
+ "loss": 5.487186431884766,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.022229891270399094,
+ "learning_rate": 0.0027815187442968995,
+ "loss": 5.488579273223877,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.026930972933769226,
+ "learning_rate": 0.002780822288802037,
+ "loss": 5.492000102996826,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.02733265422284603,
+ "learning_rate": 0.00278012481345702,
+ "loss": 5.4663801193237305,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.02993253618478775,
+ "learning_rate": 0.0027794263188237922,
+ "loss": 5.543368816375732,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03152153640985489,
+ "learning_rate": 0.002778726805465118,
+ "loss": 5.565644264221191,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.031395044177770615,
+ "learning_rate": 0.002778026273944581,
+ "loss": 5.553359508514404,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.026603225618600845,
+ "learning_rate": 0.002777324724826589,
+ "loss": 5.552347183227539,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.024152139201760292,
+ "learning_rate": 0.002776622158676366,
+ "loss": 5.493257522583008,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.602636814117432,
+ "eval_runtime": 54.6199,
+ "eval_samples_per_second": 44.709,
+ "eval_steps_per_second": 1.41,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.02089761756360531,
+ "learning_rate": 0.0027759185760599576,
+ "loss": 5.461747646331787,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.018181778490543365,
+ "learning_rate": 0.0027752139775442278,
+ "loss": 5.492741584777832,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.020166238769888878,
+ "learning_rate": 0.0027745083636968575,
+ "loss": 5.476003646850586,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.015491694211959839,
+ "learning_rate": 0.0027738017350863488,
+ "loss": 5.425764560699463,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.022291697561740875,
+ "learning_rate": 0.0027730940922820182,
+ "loss": 5.45866584777832,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.018074683845043182,
+ "learning_rate": 0.002772385435854002,
+ "loss": 5.465143203735352,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.016872268170118332,
+ "learning_rate": 0.002771675766373251,
+ "loss": 5.446003437042236,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.023103339597582817,
+ "learning_rate": 0.0027709650844115336,
+ "loss": 5.411036491394043,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.027957158163189888,
+ "learning_rate": 0.0027702533905414324,
+ "loss": 5.452442169189453,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.019935399293899536,
+ "learning_rate": 0.002769540685336348,
+ "loss": 5.412145614624023,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.014351371675729752,
+ "learning_rate": 0.0027688269693704926,
+ "loss": 5.434524059295654,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.0152254244312644,
+ "learning_rate": 0.0027681122432188948,
+ "loss": 5.428544998168945,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.015071330592036247,
+ "learning_rate": 0.0027673965074573965,
+ "loss": 5.459929466247559,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.023415958508849144,
+ "learning_rate": 0.0027666797626626537,
+ "loss": 5.447159767150879,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.017228802666068077,
+ "learning_rate": 0.0027659620094121337,
+ "loss": 5.442273139953613,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.014184504747390747,
+ "learning_rate": 0.002765243248284118,
+ "loss": 5.448305130004883,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.024384526535868645,
+ "learning_rate": 0.0027645234798576988,
+ "loss": 5.4076337814331055,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.043152231723070145,
+ "learning_rate": 0.0027638027047127813,
+ "loss": 5.45344352722168,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.02889758162200451,
+ "learning_rate": 0.002763080923430081,
+ "loss": 5.489194393157959,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.02686968818306923,
+ "learning_rate": 0.002762358136591123,
+ "loss": 5.493659496307373,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.020467868074774742,
+ "learning_rate": 0.002761634344778245,
+ "loss": 5.410489559173584,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.019759224727749825,
+ "learning_rate": 0.002760909548574592,
+ "loss": 5.439385890960693,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.0233843382447958,
+ "learning_rate": 0.0027601837485641194,
+ "loss": 5.445036888122559,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.031753458082675934,
+ "learning_rate": 0.0027594569453315916,
+ "loss": 5.503895282745361,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.02089349366724491,
+ "learning_rate": 0.002758729139462581,
+ "loss": 5.45689582824707,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.03339357674121857,
+ "learning_rate": 0.002758000331543466,
+ "loss": 5.522361755371094,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019328996539115906,
+ "learning_rate": 0.002757270522161436,
+ "loss": 5.441373825073242,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019863856956362724,
+ "learning_rate": 0.002756539711904484,
+ "loss": 5.4633283615112305,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.02141754701733589,
+ "learning_rate": 0.0027558079013614116,
+ "loss": 5.453215599060059,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016591915860772133,
+ "learning_rate": 0.002755075091121824,
+ "loss": 5.463696479797363,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.01955028623342514,
+ "learning_rate": 0.0027543412817761345,
+ "loss": 5.411355018615723,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.016791662201285362,
+ "learning_rate": 0.0027536064739155596,
+ "loss": 5.459614276885986,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.019661981612443924,
+ "learning_rate": 0.00275287066813212,
+ "loss": 5.440699577331543,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.01927965320646763,
+ "learning_rate": 0.002752133865018642,
+ "loss": 5.409865856170654,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.016483580693602562,
+ "learning_rate": 0.002751396065168754,
+ "loss": 5.390999794006348,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016264773905277252,
+ "learning_rate": 0.0027506572691768888,
+ "loss": 5.418543815612793,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.01958460919559002,
+ "learning_rate": 0.00274991747763828,
+ "loss": 5.403289794921875,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.01861695572733879,
+ "learning_rate": 0.002749176691148965,
+ "loss": 5.431911468505859,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.029594263061881065,
+ "learning_rate": 0.002748434910305782,
+ "loss": 5.399242401123047,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.028574613854289055,
+ "learning_rate": 0.0027476921357063697,
+ "loss": 5.440446853637695,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.028612585738301277,
+ "learning_rate": 0.0027469483679491693,
+ "loss": 5.410346031188965,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.020704898983240128,
+ "learning_rate": 0.00274620360763342,
+ "loss": 5.44694709777832,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.020542364567518234,
+ "learning_rate": 0.002745457855359163,
+ "loss": 5.402653694152832,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.015762662515044212,
+ "learning_rate": 0.0027447111117272363,
+ "loss": 5.396594047546387,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022096984088420868,
+ "learning_rate": 0.002743963377339278,
+ "loss": 5.398744106292725,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.018791764974594116,
+ "learning_rate": 0.002743214652797724,
+ "loss": 5.397758483886719,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01642526686191559,
+ "learning_rate": 0.0027424649387058093,
+ "loss": 5.411459922790527,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.017931703478097916,
+ "learning_rate": 0.0027417142356675637,
+ "loss": 5.390264511108398,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018819602206349373,
+ "learning_rate": 0.0027409625442878153,
+ "loss": 5.407539367675781,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.019389411434531212,
+ "learning_rate": 0.0027402098651721886,
+ "loss": 5.386474609375,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.013645983301103115,
+ "learning_rate": 0.002739456198927104,
+ "loss": 5.373845100402832,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.013846895657479763,
+ "learning_rate": 0.002738701546159776,
+ "loss": 5.387155055999756,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013553125783801079,
+ "learning_rate": 0.002737945907478215,
+ "loss": 5.377801895141602,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.012771612964570522,
+ "learning_rate": 0.002737189283491225,
+ "loss": 5.344066143035889,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.020465301349759102,
+ "learning_rate": 0.002736431674808405,
+ "loss": 5.382771968841553,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.014798106625676155,
+ "learning_rate": 0.0027356730820401465,
+ "loss": 5.370522499084473,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.01566634140908718,
+ "learning_rate": 0.0027349135057976332,
+ "loss": 5.351980209350586,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.017485102638602257,
+ "learning_rate": 0.002734152946692843,
+ "loss": 5.355443477630615,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.016646096482872963,
+ "learning_rate": 0.0027333914053385444,
+ "loss": 5.34829044342041,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.017293352633714676,
+ "learning_rate": 0.0027326288823482964,
+ "loss": 5.3519487380981445,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.018873002380132675,
+ "learning_rate": 0.002731865378336452,
+ "loss": 5.348442077636719,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.015504784882068634,
+ "learning_rate": 0.002731100893918151,
+ "loss": 5.341490268707275,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.014270870015025139,
+ "learning_rate": 0.0027303354297093254,
+ "loss": 5.359251022338867,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.018820680677890778,
+ "learning_rate": 0.002729568986326695,
+ "loss": 5.390280723571777,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.021008187904953957,
+ "learning_rate": 0.0027288015643877708,
+ "loss": 5.338380336761475,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.016266655176877975,
+ "learning_rate": 0.0027280331645108498,
+ "loss": 5.347043037414551,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.01831045374274254,
+ "learning_rate": 0.0027272637873150185,
+ "loss": 5.326537132263184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02073794975876808,
+ "learning_rate": 0.002726493433420149,
+ "loss": 5.3607025146484375,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.022534754127264023,
+ "learning_rate": 0.0027257221034469033,
+ "loss": 5.353292465209961,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.021627362817525864,
+ "learning_rate": 0.002724949798016727,
+ "loss": 5.373323917388916,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.018226243555545807,
+ "learning_rate": 0.0027241765177518523,
+ "loss": 5.344578266143799,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.02032875455915928,
+ "learning_rate": 0.002723402263275298,
+ "loss": 5.329159259796143,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.44969367980957,
+ "eval_runtime": 51.2294,
+ "eval_samples_per_second": 47.668,
+ "eval_steps_per_second": 1.503,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.01613701693713665,
+ "learning_rate": 0.0027226270352108664,
+ "loss": 5.3024678230285645,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01940973661839962,
+ "learning_rate": 0.0027218508341831445,
+ "loss": 5.270181179046631,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.015840493142604828,
+ "learning_rate": 0.002721073660817504,
+ "loss": 5.261706352233887,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.017390675842761993,
+ "learning_rate": 0.0027202955157401,
+ "loss": 5.262207984924316,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.015184012241661549,
+ "learning_rate": 0.0027195163995778686,
+ "loss": 5.297155380249023,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.013417073525488377,
+ "learning_rate": 0.00271873631295853,
+ "loss": 5.279838562011719,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.013024317100644112,
+ "learning_rate": 0.0027179552565105864,
+ "loss": 5.265446186065674,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.01426338218152523,
+ "learning_rate": 0.002717173230863321,
+ "loss": 5.284180164337158,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.019604040309786797,
+ "learning_rate": 0.002716390236646797,
+ "loss": 5.286281108856201,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.019039476290345192,
+ "learning_rate": 0.0027156062744918587,
+ "loss": 5.290433406829834,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.019284334033727646,
+ "learning_rate": 0.0027148213450301313,
+ "loss": 5.262433052062988,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.0145154669880867,
+ "learning_rate": 0.0027140354488940175,
+ "loss": 5.24387264251709,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.014776417054235935,
+ "learning_rate": 0.0027132485867166992,
+ "loss": 5.282435417175293,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.012711729854345322,
+ "learning_rate": 0.0027124607591321377,
+ "loss": 5.246735572814941,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.013394835405051708,
+ "learning_rate": 0.0027116719667750716,
+ "loss": 5.270739555358887,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.013468850404024124,
+ "learning_rate": 0.002710882210281016,
+ "loss": 5.27161169052124,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.030262840911746025,
+ "learning_rate": 0.002710091490286264,
+ "loss": 5.256673812866211,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.013560288585722446,
+ "learning_rate": 0.0027092998074278846,
+ "loss": 5.253674507141113,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014592813327908516,
+ "learning_rate": 0.002708507162343721,
+ "loss": 5.25804328918457,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014328169636428356,
+ "learning_rate": 0.0027077135556723953,
+ "loss": 5.235117435455322,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.01277366653084755,
+ "learning_rate": 0.0027069189880533,
+ "loss": 5.2053985595703125,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.012948785908520222,
+ "learning_rate": 0.0027061234601266043,
+ "loss": 5.214653015136719,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.014892828650772572,
+ "learning_rate": 0.002705326972533252,
+ "loss": 5.227547645568848,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01427230704575777,
+ "learning_rate": 0.002704529525914957,
+ "loss": 5.241065502166748,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.01650116965174675,
+ "learning_rate": 0.0027037311209142092,
+ "loss": 5.227713108062744,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.019728578627109528,
+ "learning_rate": 0.0027029317581742677,
+ "loss": 5.2624921798706055,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.023847149685025215,
+ "learning_rate": 0.0027021314383391657,
+ "loss": 5.278811454772949,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.020690469071269035,
+ "learning_rate": 0.0027013301620537056,
+ "loss": 5.238919258117676,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.021211959421634674,
+ "learning_rate": 0.0027005279299634616,
+ "loss": 5.253684997558594,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.025854922831058502,
+ "learning_rate": 0.002699724742714778,
+ "loss": 5.2538652420043945,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.0215485580265522,
+ "learning_rate": 0.0026989206009547678,
+ "loss": 5.224540710449219,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.019827792420983315,
+ "learning_rate": 0.0026981155053313137,
+ "loss": 5.220674514770508,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.017689093947410583,
+ "learning_rate": 0.002697309456493066,
+ "loss": 5.228511810302734,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.022564806044101715,
+ "learning_rate": 0.002696502455089445,
+ "loss": 5.295933723449707,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.020070280879735947,
+ "learning_rate": 0.0026956945017706357,
+ "loss": 5.230920791625977,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.025744661688804626,
+ "learning_rate": 0.002694885597187593,
+ "loss": 5.296453952789307,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.021263007074594498,
+ "learning_rate": 0.0026940757419920357,
+ "loss": 5.220481872558594,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.017577245831489563,
+ "learning_rate": 0.0026932649368364495,
+ "loss": 5.231488227844238,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.015562832355499268,
+ "learning_rate": 0.002692453182374086,
+ "loss": 5.175224304199219,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.014115403406322002,
+ "learning_rate": 0.0026916404792589606,
+ "loss": 5.228387832641602,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.012863830663263798,
+ "learning_rate": 0.002690826828145854,
+ "loss": 5.186727523803711,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.014151126146316528,
+ "learning_rate": 0.002690012229690309,
+ "loss": 5.1694817543029785,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.01316932775080204,
+ "learning_rate": 0.0026891966845486353,
+ "loss": 5.19931697845459,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.011340726166963577,
+ "learning_rate": 0.0026883801933779003,
+ "loss": 5.238619804382324,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.012616089545190334,
+ "learning_rate": 0.0026875627568359378,
+ "loss": 5.183376789093018,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.012274038046598434,
+ "learning_rate": 0.002686744375581341,
+ "loss": 5.199946403503418,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.014990398660302162,
+ "learning_rate": 0.002685925050273465,
+ "loss": 5.168451309204102,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.011399757117033005,
+ "learning_rate": 0.0026851047815724255,
+ "loss": 5.173828125,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.013638390228152275,
+ "learning_rate": 0.0026842835701390982,
+ "loss": 5.142702102661133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.014472718350589275,
+ "learning_rate": 0.0026834614166351177,
+ "loss": 5.173436164855957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.014509808272123337,
+ "learning_rate": 0.002682638321722879,
+ "loss": 5.173975467681885,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.013576583005487919,
+ "learning_rate": 0.0026818142860655352,
+ "loss": 5.1435089111328125,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.012914472259581089,
+ "learning_rate": 0.0026809893103269954,
+ "loss": 5.151381492614746,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.013803352601826191,
+ "learning_rate": 0.0026801633951719288,
+ "loss": 5.133159637451172,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.018488213419914246,
+ "learning_rate": 0.0026793365412657606,
+ "loss": 5.183903694152832,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.01843969337642193,
+ "learning_rate": 0.002678508749274671,
+ "loss": 5.1660356521606445,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.017181839793920517,
+ "learning_rate": 0.0026776800198655983,
+ "loss": 5.179847717285156,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0148160420358181,
+ "learning_rate": 0.0026768503537062333,
+ "loss": 5.149198532104492,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.013251296244561672,
+ "learning_rate": 0.0026760197514650247,
+ "loss": 5.192000389099121,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.010746927931904793,
+ "learning_rate": 0.0026751882138111717,
+ "loss": 5.138680458068848,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.012316994369029999,
+ "learning_rate": 0.002674355741414631,
+ "loss": 5.169965744018555,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.01191797386854887,
+ "learning_rate": 0.0026735223349461106,
+ "loss": 5.128267288208008,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.013525084592401981,
+ "learning_rate": 0.0026726879950770698,
+ "loss": 5.124795436859131,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.014843370765447617,
+ "learning_rate": 0.002671852722479721,
+ "loss": 5.1497087478637695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.015285412780940533,
+ "learning_rate": 0.00267101651782703,
+ "loss": 5.093118190765381,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.013555832207202911,
+ "learning_rate": 0.0026701793817927107,
+ "loss": 5.13249397277832,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.010569742880761623,
+ "learning_rate": 0.0026693413150512273,
+ "loss": 5.084077835083008,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.011483813636004925,
+ "learning_rate": 0.0026685023182777975,
+ "loss": 5.128049850463867,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.013724366202950478,
+ "learning_rate": 0.0026676623921483844,
+ "loss": 5.157539367675781,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.014337432570755482,
+ "learning_rate": 0.0026668215373397016,
+ "loss": 5.149948596954346,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.014101454056799412,
+ "learning_rate": 0.00266597975452921,
+ "loss": 5.117250442504883,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.016004368662834167,
+ "learning_rate": 0.0026651370443951197,
+ "loss": 5.143221855163574,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.249833583831787,
+ "eval_runtime": 51.5543,
+ "eval_samples_per_second": 47.368,
+ "eval_steps_per_second": 1.494,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.01297403872013092,
+ "learning_rate": 0.0026642934076163867,
+ "loss": 5.089287281036377,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.012867718935012817,
+ "learning_rate": 0.0026634488448727144,
+ "loss": 5.0784101486206055,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.014600927010178566,
+ "learning_rate": 0.0026626033568445504,
+ "loss": 5.050220489501953,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.016619572415947914,
+ "learning_rate": 0.0026617569442130905,
+ "loss": 5.079303741455078,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.015560587868094444,
+ "learning_rate": 0.0026609096076602736,
+ "loss": 5.072385311126709,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.01774057187139988,
+ "learning_rate": 0.0026600613478687833,
+ "loss": 5.122529983520508,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.019971400499343872,
+ "learning_rate": 0.0026592121655220472,
+ "loss": 5.085966110229492,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.017667256295681,
+ "learning_rate": 0.0026583620613042365,
+ "loss": 5.078432559967041,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.022369660437107086,
+ "learning_rate": 0.002657511035900264,
+ "loss": 5.0630903244018555,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.017964458093047142,
+ "learning_rate": 0.002656659089995787,
+ "loss": 5.0846076011657715,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.021769145503640175,
+ "learning_rate": 0.002655806224277201,
+ "loss": 5.081344127655029,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.02487620897591114,
+ "learning_rate": 0.0026549524394316464,
+ "loss": 5.106287956237793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.022516218945384026,
+ "learning_rate": 0.0026540977361470012,
+ "loss": 5.128098487854004,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.02099926583468914,
+ "learning_rate": 0.0026532421151118845,
+ "loss": 5.103562355041504,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.01899387687444687,
+ "learning_rate": 0.002652385577015655,
+ "loss": 5.062675476074219,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.016801996156573296,
+ "learning_rate": 0.002651528122548409,
+ "loss": 5.096002578735352,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.014009200967848301,
+ "learning_rate": 0.0026506697524009836,
+ "loss": 5.075201034545898,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.014653858728706837,
+ "learning_rate": 0.002649810467264951,
+ "loss": 5.090478897094727,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.014918150380253792,
+ "learning_rate": 0.0026489502678326226,
+ "loss": 5.070797920227051,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.01327288057655096,
+ "learning_rate": 0.0026480891547970443,
+ "loss": 5.030500411987305,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.012885140255093575,
+ "learning_rate": 0.002647227128852,
+ "loss": 5.0845537185668945,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.012848512269556522,
+ "learning_rate": 0.002646364190692009,
+ "loss": 5.079484939575195,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.01149789709597826,
+ "learning_rate": 0.002645500341012324,
+ "loss": 5.036316394805908,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.010360757820308208,
+ "learning_rate": 0.0026446355805089337,
+ "loss": 5.04129695892334,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.010426290333271027,
+ "learning_rate": 0.00264376990987856,
+ "loss": 5.051270484924316,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.01018153689801693,
+ "learning_rate": 0.002642903329818658,
+ "loss": 5.102342128753662,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.011879283003509045,
+ "learning_rate": 0.0026420358410274153,
+ "loss": 5.053501129150391,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.010418329387903214,
+ "learning_rate": 0.002641167444203752,
+ "loss": 5.030653476715088,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.012302013114094734,
+ "learning_rate": 0.0026402981400473202,
+ "loss": 5.0402069091796875,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.011392916552722454,
+ "learning_rate": 0.002639427929258502,
+ "loss": 5.024111747741699,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.013075058348476887,
+ "learning_rate": 0.0026385568125384114,
+ "loss": 5.021731376647949,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015349832363426685,
+ "learning_rate": 0.0026376847905888903,
+ "loss": 5.03192138671875,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01588309183716774,
+ "learning_rate": 0.0026368118641125113,
+ "loss": 4.966363906860352,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.014378590509295464,
+ "learning_rate": 0.0026359380338125764,
+ "loss": 5.070773124694824,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.012306655757129192,
+ "learning_rate": 0.002635063300393114,
+ "loss": 5.025704860687256,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01212004292756319,
+ "learning_rate": 0.002634187664558881,
+ "loss": 4.965104103088379,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013538849540054798,
+ "learning_rate": 0.0026333111270153624,
+ "loss": 5.029810905456543,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.012052906677126884,
+ "learning_rate": 0.0026324336884687672,
+ "loss": 5.0262908935546875,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.011368660256266594,
+ "learning_rate": 0.002631555349626033,
+ "loss": 5.001485824584961,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.010215395130217075,
+ "learning_rate": 0.0026306761111948205,
+ "loss": 5.019739151000977,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.011449878104031086,
+ "learning_rate": 0.0026297959738835167,
+ "loss": 4.9816083908081055,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.012381811626255512,
+ "learning_rate": 0.0026289149384012326,
+ "loss": 5.025032043457031,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.00974641740322113,
+ "learning_rate": 0.002628033005457802,
+ "loss": 5.000177383422852,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.009515874087810516,
+ "learning_rate": 0.0026271501757637833,
+ "loss": 4.996979713439941,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.011125455610454082,
+ "learning_rate": 0.0026262664500304555,
+ "loss": 4.996796607971191,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.013290468603372574,
+ "learning_rate": 0.0026253818289698207,
+ "loss": 5.053703308105469,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.014650283381342888,
+ "learning_rate": 0.002624496313294602,
+ "loss": 5.001321792602539,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014537827111780643,
+ "learning_rate": 0.002623609903718244,
+ "loss": 5.024584770202637,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.012908566743135452,
+ "learning_rate": 0.0026227226009549105,
+ "loss": 5.042878150939941,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.01446808222681284,
+ "learning_rate": 0.002621834405719485,
+ "loss": 5.020857810974121,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.01535153016448021,
+ "learning_rate": 0.0026209453187275708,
+ "loss": 5.0056304931640625,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01815616339445114,
+ "learning_rate": 0.0026200553406954892,
+ "loss": 5.0467376708984375,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.023135516792535782,
+ "learning_rate": 0.0026191644723402795,
+ "loss": 5.033995628356934,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.019682930782437325,
+ "learning_rate": 0.002618272714379698,
+ "loss": 5.008149147033691,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.018521904945373535,
+ "learning_rate": 0.002617380067532217,
+ "loss": 5.007076263427734,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.02118663489818573,
+ "learning_rate": 0.0026164865325170278,
+ "loss": 5.05519962310791,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024684788659214973,
+ "learning_rate": 0.0026155921100540344,
+ "loss": 5.027571201324463,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.02609819918870926,
+ "learning_rate": 0.0026146968008638567,
+ "loss": 5.10361385345459,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.025958053767681122,
+ "learning_rate": 0.00261380060566783,
+ "loss": 5.0911455154418945,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.017067518085241318,
+ "learning_rate": 0.002612903525188002,
+ "loss": 5.045783042907715,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.016510486602783203,
+ "learning_rate": 0.0026120055601471345,
+ "loss": 5.069464683532715,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015552333556115627,
+ "learning_rate": 0.0026111067112687012,
+ "loss": 5.024248123168945,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.013432685285806656,
+ "learning_rate": 0.0026102069792768886,
+ "loss": 5.004837989807129,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.012300201691687107,
+ "learning_rate": 0.0026093063648965943,
+ "loss": 5.031093597412109,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.01197744719684124,
+ "learning_rate": 0.002608404868853428,
+ "loss": 4.965237617492676,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.01100246049463749,
+ "learning_rate": 0.002607502491873708,
+ "loss": 5.04719352722168,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.010111447423696518,
+ "learning_rate": 0.0026065992346844626,
+ "loss": 4.986410617828369,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.010669790208339691,
+ "learning_rate": 0.002605695098013431,
+ "loss": 4.9825663566589355,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.010805477388203144,
+ "learning_rate": 0.0026047900825890593,
+ "loss": 4.991506576538086,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.010886074975132942,
+ "learning_rate": 0.002603884189140502,
+ "loss": 4.987531661987305,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.010735839605331421,
+ "learning_rate": 0.0026029774183976206,
+ "loss": 4.982756614685059,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.011092593893408775,
+ "learning_rate": 0.002602069771090985,
+ "loss": 4.994824409484863,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 5.13364315032959,
+ "eval_runtime": 50.4195,
+ "eval_samples_per_second": 48.434,
+ "eval_steps_per_second": 1.527,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.01005905494093895,
+ "learning_rate": 0.0026011612479518694,
+ "loss": 4.918412685394287,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.009904682636260986,
+ "learning_rate": 0.002600251849712255,
+ "loss": 4.924630165100098,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.009543103165924549,
+ "learning_rate": 0.0025993415771048273,
+ "loss": 4.940309524536133,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.011031470261514187,
+ "learning_rate": 0.0025984304308629765,
+ "loss": 4.924609184265137,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.009511022828519344,
+ "learning_rate": 0.002597518411720796,
+ "loss": 4.9102325439453125,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.00886621605604887,
+ "learning_rate": 0.0025966055204130847,
+ "loss": 4.9236907958984375,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.010057039558887482,
+ "learning_rate": 0.002595691757675341,
+ "loss": 4.908633232116699,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.01098555512726307,
+ "learning_rate": 0.0025947771242437684,
+ "loss": 4.935927391052246,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.008800056762993336,
+ "learning_rate": 0.0025938616208552694,
+ "loss": 4.873291969299316,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.009747766889631748,
+ "learning_rate": 0.0025929452482474497,
+ "loss": 4.954436302185059,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.010785684920847416,
+ "learning_rate": 0.002592028007158613,
+ "loss": 4.924927711486816,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.012620717287063599,
+ "learning_rate": 0.0025911098983277645,
+ "loss": 4.953606605529785,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.014428738504648209,
+ "learning_rate": 0.002590190922494607,
+ "loss": 4.951706886291504,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.014496048912405968,
+ "learning_rate": 0.0025892710803995444,
+ "loss": 4.941249847412109,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.0109238987788558,
+ "learning_rate": 0.0025883503727836764,
+ "loss": 4.939975261688232,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.012271072715520859,
+ "learning_rate": 0.0025874288003887996,
+ "loss": 4.880942344665527,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.016302887350320816,
+ "learning_rate": 0.0025865063639574086,
+ "loss": 4.931602478027344,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.016137251630425453,
+ "learning_rate": 0.0025855830642326945,
+ "loss": 4.918013095855713,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.014732961542904377,
+ "learning_rate": 0.002584658901958542,
+ "loss": 4.927577495574951,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.016387946903705597,
+ "learning_rate": 0.0025837338778795325,
+ "loss": 4.930819034576416,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.017146151512861252,
+ "learning_rate": 0.002582807992740941,
+ "loss": 4.928377151489258,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.016153613105416298,
+ "learning_rate": 0.0025818812472887375,
+ "loss": 4.975607395172119,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.013386047445237637,
+ "learning_rate": 0.0025809536422695825,
+ "loss": 4.9434709548950195,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.012328253127634525,
+ "learning_rate": 0.0025800251784308314,
+ "loss": 4.917585372924805,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.012000586837530136,
+ "learning_rate": 0.0025790958565205316,
+ "loss": 4.9137725830078125,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.009463997557759285,
+ "learning_rate": 0.002578165677287419,
+ "loss": 4.910362243652344,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.010591360740363598,
+ "learning_rate": 0.0025772346414809242,
+ "loss": 4.916032314300537,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.010429762303829193,
+ "learning_rate": 0.002576302749851164,
+ "loss": 4.934561729431152,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.12526583671569824,
+ "learning_rate": 0.0025753700031489486,
+ "loss": 4.897515773773193,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01368892751634121,
+ "learning_rate": 0.0025744364021257743,
+ "loss": 4.94814920425415,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.016603967174887657,
+ "learning_rate": 0.0025735019475338254,
+ "loss": 4.8996758460998535,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.06110106781125069,
+ "learning_rate": 0.0025725666401259775,
+ "loss": 4.946794509887695,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.018292423337697983,
+ "learning_rate": 0.002571630480655789,
+ "loss": 4.958992958068848,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.01852245442569256,
+ "learning_rate": 0.0025706934698775073,
+ "loss": 4.942556381225586,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.0195649154484272,
+ "learning_rate": 0.0025697556085460646,
+ "loss": 4.895773887634277,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.017820244655013084,
+ "learning_rate": 0.00256881689741708,
+ "loss": 4.915250778198242,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.016814742237329483,
+ "learning_rate": 0.002567877337246855,
+ "loss": 4.9521260261535645,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.015319781377911568,
+ "learning_rate": 0.002566936928792377,
+ "loss": 4.899402141571045,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.01607401855289936,
+ "learning_rate": 0.0025659956728113146,
+ "loss": 4.909913063049316,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.015564955770969391,
+ "learning_rate": 0.002565053570062023,
+ "loss": 4.947719573974609,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.017502037808299065,
+ "learning_rate": 0.002564110621303536,
+ "loss": 4.95107364654541,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.014791073277592659,
+ "learning_rate": 0.002563166827295571,
+ "loss": 4.8995866775512695,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.012700124643743038,
+ "learning_rate": 0.002562222188798525,
+ "loss": 4.922100067138672,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.011235272511839867,
+ "learning_rate": 0.0025612767065734763,
+ "loss": 4.884456634521484,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.013009768910706043,
+ "learning_rate": 0.002560330381382183,
+ "loss": 4.869503021240234,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.0102638378739357,
+ "learning_rate": 0.0025593832139870834,
+ "loss": 4.934551239013672,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.011081870645284653,
+ "learning_rate": 0.0025584352051512913,
+ "loss": 4.937268257141113,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.010681218467652798,
+ "learning_rate": 0.002557486355638601,
+ "loss": 4.920559883117676,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.010962890461087227,
+ "learning_rate": 0.002556536666213484,
+ "loss": 4.936082363128662,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.010790379717946053,
+ "learning_rate": 0.0025555861376410868,
+ "loss": 4.9506449699401855,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.010659157298505306,
+ "learning_rate": 0.0025546347706872346,
+ "loss": 4.872883319854736,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.012962655164301395,
+ "learning_rate": 0.0025536825661184254,
+ "loss": 4.901210784912109,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.014810293912887573,
+ "learning_rate": 0.002552729524701833,
+ "loss": 4.895804405212402,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.010993832722306252,
+ "learning_rate": 0.0025517756472053068,
+ "loss": 4.891035079956055,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.013413915410637856,
+ "learning_rate": 0.002550820934397368,
+ "loss": 4.900317192077637,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.01580147072672844,
+ "learning_rate": 0.0025498653870472117,
+ "loss": 4.883669853210449,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01864512264728546,
+ "learning_rate": 0.0025489090059247045,
+ "loss": 4.881938457489014,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017185743898153305,
+ "learning_rate": 0.0025479517918003862,
+ "loss": 4.853325843811035,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.012045156210660934,
+ "learning_rate": 0.002546993745445466,
+ "loss": 4.9152326583862305,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.013420728035271168,
+ "learning_rate": 0.0025460348676318257,
+ "loss": 4.895130157470703,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.011561610735952854,
+ "learning_rate": 0.0025450751591320152,
+ "loss": 4.890257835388184,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013164008036255836,
+ "learning_rate": 0.0025441146207192547,
+ "loss": 4.885972023010254,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01120204571634531,
+ "learning_rate": 0.002543153253167432,
+ "loss": 4.882246971130371,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.011342627927660942,
+ "learning_rate": 0.002542191057251104,
+ "loss": 4.862927436828613,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.010852213948965073,
+ "learning_rate": 0.0025412280337454942,
+ "loss": 4.879744529724121,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.012851282022893429,
+ "learning_rate": 0.0025402641834264936,
+ "loss": 4.882238388061523,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.011839655227959156,
+ "learning_rate": 0.002539299507070659,
+ "loss": 4.85658073425293,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.011315262876451015,
+ "learning_rate": 0.002538334005455213,
+ "loss": 4.9234771728515625,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.011661105789244175,
+ "learning_rate": 0.0025373676793580416,
+ "loss": 4.855696678161621,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.009712521918118,
+ "learning_rate": 0.0025364005295576975,
+ "loss": 4.89942741394043,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.011002928018569946,
+ "learning_rate": 0.0025354325568333956,
+ "loss": 4.871459007263184,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.01509513333439827,
+ "learning_rate": 0.0025344637619650135,
+ "loss": 4.834778785705566,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 5.031810283660889,
+ "eval_runtime": 51.9412,
+ "eval_samples_per_second": 47.015,
+ "eval_steps_per_second": 1.482,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.010317740961909294,
+ "learning_rate": 0.0025334941457330925,
+ "loss": 4.8169708251953125,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.00892715621739626,
+ "learning_rate": 0.0025325237089188343,
+ "loss": 4.852178573608398,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.011608307249844074,
+ "learning_rate": 0.002531552452304103,
+ "loss": 4.790487289428711,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.013100114651024342,
+ "learning_rate": 0.0025305803766714216,
+ "loss": 4.863536834716797,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.016571057960391045,
+ "learning_rate": 0.002529607482803974,
+ "loss": 4.840987205505371,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.014595622196793556,
+ "learning_rate": 0.002528633771485604,
+ "loss": 4.815027236938477,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.014603931456804276,
+ "learning_rate": 0.002527659243500812,
+ "loss": 4.853024959564209,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.019107041880488396,
+ "learning_rate": 0.0025266838996347595,
+ "loss": 4.829808235168457,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.020292190834879875,
+ "learning_rate": 0.0025257077406732613,
+ "loss": 4.775232315063477,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.014630074612796307,
+ "learning_rate": 0.0025247307674027926,
+ "loss": 4.826682090759277,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.019507668912410736,
+ "learning_rate": 0.0025237529806104823,
+ "loss": 4.843222141265869,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.023950330913066864,
+ "learning_rate": 0.002522774381084116,
+ "loss": 4.820512771606445,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.022547820582985878,
+ "learning_rate": 0.002521794969612133,
+ "loss": 4.854910850524902,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.017819277942180634,
+ "learning_rate": 0.0025208147469836283,
+ "loss": 4.825885772705078,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.017989592626690865,
+ "learning_rate": 0.002519833713988349,
+ "loss": 4.8357014656066895,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.013516640290617943,
+ "learning_rate": 0.0025188518714166956,
+ "loss": 4.83729362487793,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.015956101939082146,
+ "learning_rate": 0.002517869220059721,
+ "loss": 4.807845592498779,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.013967473991215229,
+ "learning_rate": 0.002516885760709131,
+ "loss": 4.828944206237793,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.01270931027829647,
+ "learning_rate": 0.0025159014941572785,
+ "loss": 4.826594352722168,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.012433716095983982,
+ "learning_rate": 0.002514916421197171,
+ "loss": 4.797939777374268,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.011882808059453964,
+ "learning_rate": 0.0025139305426224636,
+ "loss": 4.78917932510376,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.011193005368113518,
+ "learning_rate": 0.0025129438592274614,
+ "loss": 4.787964820861816,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.011873407289385796,
+ "learning_rate": 0.002511956371807116,
+ "loss": 4.81754732131958,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.011407418176531792,
+ "learning_rate": 0.002510968081157029,
+ "loss": 4.781639575958252,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.009560304693877697,
+ "learning_rate": 0.002509978988073449,
+ "loss": 4.8163957595825195,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.009969786740839481,
+ "learning_rate": 0.0025089890933532697,
+ "loss": 4.795045852661133,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.010739767923951149,
+ "learning_rate": 0.0025079983977940312,
+ "loss": 4.847584247589111,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.009999158792197704,
+ "learning_rate": 0.0025070069021939194,
+ "loss": 4.834909439086914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.009302482940256596,
+ "learning_rate": 0.002506014607351765,
+ "loss": 4.808745384216309,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.010687744244933128,
+ "learning_rate": 0.0025050215140670403,
+ "loss": 4.789101600646973,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.010930616408586502,
+ "learning_rate": 0.0025040276231398637,
+ "loss": 4.817464351654053,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.013101065531373024,
+ "learning_rate": 0.0025030329353709956,
+ "loss": 4.827780723571777,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.01223307941108942,
+ "learning_rate": 0.0025020374515618373,
+ "loss": 4.8367414474487305,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.013091634958982468,
+ "learning_rate": 0.0025010411725144323,
+ "loss": 4.839454650878906,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.012114843353629112,
+ "learning_rate": 0.0025000440990314643,
+ "loss": 4.832447052001953,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014664588496088982,
+ "learning_rate": 0.0024990462319162588,
+ "loss": 4.740812301635742,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.015566283836960793,
+ "learning_rate": 0.002498047571972777,
+ "loss": 4.815671920776367,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.012440434657037258,
+ "learning_rate": 0.002497048120005623,
+ "loss": 4.769775867462158,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01175499614328146,
+ "learning_rate": 0.0024960478768200362,
+ "loss": 4.8101348876953125,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.012699508108198643,
+ "learning_rate": 0.002495046843221896,
+ "loss": 4.827281951904297,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.012438303790986538,
+ "learning_rate": 0.002494045020017715,
+ "loss": 4.786289691925049,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.01709924265742302,
+ "learning_rate": 0.002493042408014646,
+ "loss": 4.828742027282715,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.022684535011649132,
+ "learning_rate": 0.0024920390080204733,
+ "loss": 4.822258949279785,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.011984334327280521,
+ "learning_rate": 0.00249103482084362,
+ "loss": 4.792006969451904,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.013941102661192417,
+ "learning_rate": 0.00249002984729314,
+ "loss": 4.7926177978515625,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.0145634226500988,
+ "learning_rate": 0.002489024088178724,
+ "loss": 4.819155216217041,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.013391612097620964,
+ "learning_rate": 0.0024880175443106932,
+ "loss": 4.777242660522461,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01764756441116333,
+ "learning_rate": 0.0024870102165000015,
+ "loss": 4.81627893447876,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.015397441573441029,
+ "learning_rate": 0.002486002105558235,
+ "loss": 4.869495391845703,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01225620973855257,
+ "learning_rate": 0.0024849932122976105,
+ "loss": 4.784395217895508,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.01534759160131216,
+ "learning_rate": 0.002483983537530975,
+ "loss": 4.813475608825684,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.015057779848575592,
+ "learning_rate": 0.002482973082071805,
+ "loss": 4.8281569480896,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.01243986189365387,
+ "learning_rate": 0.0024819618467342066,
+ "loss": 4.7866973876953125,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.013326527550816536,
+ "learning_rate": 0.0024809498323329136,
+ "loss": 4.750214576721191,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.013644695281982422,
+ "learning_rate": 0.002479937039683288,
+ "loss": 4.790144920349121,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01296213548630476,
+ "learning_rate": 0.002478923469601319,
+ "loss": 4.812170028686523,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.011848201975226402,
+ "learning_rate": 0.0024779091229036204,
+ "loss": 4.798855781555176,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.012765968218445778,
+ "learning_rate": 0.0024768940004074335,
+ "loss": 4.8272294998168945,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.011700666509568691,
+ "learning_rate": 0.002475878102930625,
+ "loss": 4.762265205383301,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.01018498931080103,
+ "learning_rate": 0.0024748614312916843,
+ "loss": 4.796628952026367,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.010241681709885597,
+ "learning_rate": 0.002473843986309725,
+ "loss": 4.829319000244141,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.009332080371677876,
+ "learning_rate": 0.0024728257688044855,
+ "loss": 4.783773899078369,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.012860726565122604,
+ "learning_rate": 0.002471806779596324,
+ "loss": 4.802700519561768,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.011156697757542133,
+ "learning_rate": 0.0024707870195062225,
+ "loss": 4.808038234710693,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.010329720564186573,
+ "learning_rate": 0.002469766489355783,
+ "loss": 4.761645317077637,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.011093895882368088,
+ "learning_rate": 0.0024687451899672278,
+ "loss": 4.795220851898193,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.010740782134234905,
+ "learning_rate": 0.0024677231221634,
+ "loss": 4.79289436340332,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.012272508814930916,
+ "learning_rate": 0.0024667002867677606,
+ "loss": 4.811873435974121,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.008665879257023335,
+ "learning_rate": 0.0024656766846043905,
+ "loss": 4.744179725646973,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.00933373998850584,
+ "learning_rate": 0.0024646523164979866,
+ "loss": 4.805453777313232,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.010399337857961655,
+ "learning_rate": 0.002463627183273864,
+ "loss": 4.75568962097168,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.013210986740887165,
+ "learning_rate": 0.0024626012857579542,
+ "loss": 4.797974586486816,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.967026233673096,
+ "eval_runtime": 51.07,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.01731858029961586,
+ "learning_rate": 0.0024615746247768043,
+ "loss": 4.748838901519775,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.018902689218521118,
+ "learning_rate": 0.0024605472011575763,
+ "loss": 4.717338562011719,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.01734415628015995,
+ "learning_rate": 0.0024595190157280468,
+ "loss": 4.737737655639648,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.01971321739256382,
+ "learning_rate": 0.002458490069316607,
+ "loss": 4.774558067321777,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.020029844716191292,
+ "learning_rate": 0.002457460362752259,
+ "loss": 4.739579200744629,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.020853109657764435,
+ "learning_rate": 0.0024564298968646203,
+ "loss": 4.791436672210693,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.019159914925694466,
+ "learning_rate": 0.0024553986724839176,
+ "loss": 4.765683650970459,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.0165473110973835,
+ "learning_rate": 0.0024543666904409903,
+ "loss": 4.732603073120117,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.017426826059818268,
+ "learning_rate": 0.002453333951567287,
+ "loss": 4.7360639572143555,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014590734615921974,
+ "learning_rate": 0.002452300456694868,
+ "loss": 4.750031471252441,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.011172222904860973,
+ "learning_rate": 0.0024512662066563996,
+ "loss": 4.740771293640137,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.012294141575694084,
+ "learning_rate": 0.0024502312022851606,
+ "loss": 4.726214408874512,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.01146282535046339,
+ "learning_rate": 0.002449195444415034,
+ "loss": 4.753599166870117,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.011197247542440891,
+ "learning_rate": 0.0024481589338805113,
+ "loss": 4.751479148864746,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.013252820819616318,
+ "learning_rate": 0.0024471216715166904,
+ "loss": 4.743409156799316,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.012366211041808128,
+ "learning_rate": 0.0024460836581592744,
+ "loss": 4.731625556945801,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.013906655833125114,
+ "learning_rate": 0.0024450448946445737,
+ "loss": 4.768153190612793,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.012118754908442497,
+ "learning_rate": 0.002444005381809499,
+ "loss": 4.74198579788208,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.01197550818324089,
+ "learning_rate": 0.002442965120491568,
+ "loss": 4.737289905548096,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.010640372522175312,
+ "learning_rate": 0.002441924111528901,
+ "loss": 4.7472686767578125,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.013553396798670292,
+ "learning_rate": 0.0024408823557602196,
+ "loss": 4.711215972900391,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.01393711194396019,
+ "learning_rate": 0.002439839854024847,
+ "loss": 4.720060348510742,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01589774526655674,
+ "learning_rate": 0.002438796607162709,
+ "loss": 4.695697784423828,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.0097908740863204,
+ "learning_rate": 0.0024377526160143303,
+ "loss": 4.729333877563477,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.011745489202439785,
+ "learning_rate": 0.0024367078814208363,
+ "loss": 4.722957134246826,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.009649662300944328,
+ "learning_rate": 0.0024356624042239498,
+ "loss": 4.727540493011475,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.008429240435361862,
+ "learning_rate": 0.002434616185265993,
+ "loss": 4.705052375793457,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.00984937883913517,
+ "learning_rate": 0.002433569225389887,
+ "loss": 4.721161365509033,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.009237290360033512,
+ "learning_rate": 0.0024325215254391466,
+ "loss": 4.724277496337891,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.008900545537471771,
+ "learning_rate": 0.002431473086257886,
+ "loss": 4.711522102355957,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.010003026574850082,
+ "learning_rate": 0.002430423908690813,
+ "loss": 4.728816032409668,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.011658324860036373,
+ "learning_rate": 0.0024293739935832316,
+ "loss": 4.716179847717285,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.01312185451388359,
+ "learning_rate": 0.0024283233417810393,
+ "loss": 4.7172346115112305,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.014306646771728992,
+ "learning_rate": 0.0024272719541307266,
+ "loss": 4.704054832458496,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.009844634681940079,
+ "learning_rate": 0.0024262198314793783,
+ "loss": 4.72379207611084,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.00903896801173687,
+ "learning_rate": 0.0024251669746746703,
+ "loss": 4.711025238037109,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.010476433672010899,
+ "learning_rate": 0.0024241133845648703,
+ "loss": 4.711664199829102,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.009128391742706299,
+ "learning_rate": 0.0024230590619988375,
+ "loss": 4.72395133972168,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.01228832732886076,
+ "learning_rate": 0.0024220040078260195,
+ "loss": 4.747347831726074,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.011916602961719036,
+ "learning_rate": 0.002420948222896455,
+ "loss": 4.696471691131592,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.012877700850367546,
+ "learning_rate": 0.0024198917080607706,
+ "loss": 4.731119632720947,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.012967764399945736,
+ "learning_rate": 0.0024188344641701814,
+ "loss": 4.688953876495361,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.011993546038866043,
+ "learning_rate": 0.002417776492076489,
+ "loss": 4.695237159729004,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.011188571341335773,
+ "learning_rate": 0.0024167177926320833,
+ "loss": 4.699850082397461,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.010556348599493504,
+ "learning_rate": 0.0024156583666899395,
+ "loss": 4.703063011169434,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.012023619376122952,
+ "learning_rate": 0.0024145982151036176,
+ "loss": 4.6918792724609375,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01266001258045435,
+ "learning_rate": 0.0024135373387272623,
+ "loss": 4.6653289794921875,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.012152725830674171,
+ "learning_rate": 0.0024124757384156024,
+ "loss": 4.682897567749023,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.014056944288313389,
+ "learning_rate": 0.002411413415023951,
+ "loss": 4.696136474609375,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.016826556995511055,
+ "learning_rate": 0.002410350369408202,
+ "loss": 4.701657295227051,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.015048800967633724,
+ "learning_rate": 0.0024092866024248323,
+ "loss": 4.691389083862305,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.012512968853116035,
+ "learning_rate": 0.0024082221149308997,
+ "loss": 4.677886009216309,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01596425101161003,
+ "learning_rate": 0.0024071569077840425,
+ "loss": 4.718337535858154,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01640196517109871,
+ "learning_rate": 0.002406090981842479,
+ "loss": 4.697303295135498,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.012347170151770115,
+ "learning_rate": 0.002405024337965007,
+ "loss": 4.696938991546631,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01213028747588396,
+ "learning_rate": 0.0024039569770110012,
+ "loss": 4.737306594848633,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01314542256295681,
+ "learning_rate": 0.002402888899840416,
+ "loss": 4.69948673248291,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.013764189556241035,
+ "learning_rate": 0.0024018201073137805,
+ "loss": 4.722800254821777,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.011147351004183292,
+ "learning_rate": 0.002400750600292203,
+ "loss": 4.694892883300781,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.015437614172697067,
+ "learning_rate": 0.0023996803796373653,
+ "loss": 4.704752445220947,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.016402097418904305,
+ "learning_rate": 0.0023986094462115255,
+ "loss": 4.67849063873291,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.011600594967603683,
+ "learning_rate": 0.002397537800877514,
+ "loss": 4.689641952514648,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.011864298023283482,
+ "learning_rate": 0.0023964654444987373,
+ "loss": 4.670377731323242,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01409872155636549,
+ "learning_rate": 0.002395392377939173,
+ "loss": 4.696022033691406,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.012106554582715034,
+ "learning_rate": 0.0023943186020633717,
+ "loss": 4.705615043640137,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.012310667894780636,
+ "learning_rate": 0.002393244117736455,
+ "loss": 4.702986717224121,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.013501073233783245,
+ "learning_rate": 0.002392168925824116,
+ "loss": 4.6700215339660645,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.013565126806497574,
+ "learning_rate": 0.0023910930271926166,
+ "loss": 4.699914455413818,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.013421077281236649,
+ "learning_rate": 0.00239001642270879,
+ "loss": 4.713191986083984,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.014408102259039879,
+ "learning_rate": 0.002388939113240035,
+ "loss": 4.655015468597412,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.013502835296094418,
+ "learning_rate": 0.0023878610996543224,
+ "loss": 4.6142258644104,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.014169570058584213,
+ "learning_rate": 0.002386782382820188,
+ "loss": 4.660243034362793,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.885881423950195,
+ "eval_runtime": 52.0315,
+ "eval_samples_per_second": 46.933,
+ "eval_steps_per_second": 1.48,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.014618083834648132,
+ "learning_rate": 0.002385702963606734,
+ "loss": 4.640046119689941,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01745622046291828,
+ "learning_rate": 0.0023846228428836278,
+ "loss": 4.641966819763184,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.014995258301496506,
+ "learning_rate": 0.002383542021521106,
+ "loss": 4.60922908782959,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.013278020545840263,
+ "learning_rate": 0.002382460500389964,
+ "loss": 4.652071952819824,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.014685257337987423,
+ "learning_rate": 0.002381378280361566,
+ "loss": 4.672174453735352,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.012737024575471878,
+ "learning_rate": 0.002380295362307836,
+ "loss": 4.612064361572266,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.012992842122912407,
+ "learning_rate": 0.002379211747101261,
+ "loss": 4.611995697021484,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.010210646316409111,
+ "learning_rate": 0.0023781274356148912,
+ "loss": 4.6055707931518555,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.010265499353408813,
+ "learning_rate": 0.0023770424287223365,
+ "loss": 4.637089729309082,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.010174009948968887,
+ "learning_rate": 0.002375956727297768,
+ "loss": 4.620571136474609,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.009512446820735931,
+ "learning_rate": 0.0023748703322159143,
+ "loss": 4.644092559814453,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.009849252179265022,
+ "learning_rate": 0.0023737832443520655,
+ "loss": 4.625055313110352,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.010493188165128231,
+ "learning_rate": 0.0023726954645820674,
+ "loss": 4.644410610198975,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.022181199863553047,
+ "learning_rate": 0.002371606993782326,
+ "loss": 4.593084335327148,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.011906919069588184,
+ "learning_rate": 0.0023705178328298016,
+ "loss": 4.6452836990356445,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.011874495074152946,
+ "learning_rate": 0.0023694279826020122,
+ "loss": 4.599322319030762,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.00989708211272955,
+ "learning_rate": 0.0023683374439770295,
+ "loss": 4.607342720031738,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.009256849065423012,
+ "learning_rate": 0.0023672462178334817,
+ "loss": 4.621583461761475,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.013187132775783539,
+ "learning_rate": 0.0023661543050505496,
+ "loss": 4.614629745483398,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.013505950570106506,
+ "learning_rate": 0.002365061706507968,
+ "loss": 4.629693984985352,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.012032733298838139,
+ "learning_rate": 0.0023639684230860225,
+ "loss": 4.671209335327148,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.01278337650001049,
+ "learning_rate": 0.002362874455665553,
+ "loss": 4.648874759674072,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.01123690977692604,
+ "learning_rate": 0.0023617798051279496,
+ "loss": 4.611470699310303,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.012564311735332012,
+ "learning_rate": 0.002360684472355151,
+ "loss": 4.659026145935059,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.01206638477742672,
+ "learning_rate": 0.002359588458229649,
+ "loss": 4.642566204071045,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.010300020687282085,
+ "learning_rate": 0.0023584917636344806,
+ "loss": 4.629585266113281,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.0102265365421772,
+ "learning_rate": 0.0023573943894532336,
+ "loss": 4.609250068664551,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.012440572492778301,
+ "learning_rate": 0.002356296336570042,
+ "loss": 4.636714458465576,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.011830241419374943,
+ "learning_rate": 0.0023551976058695887,
+ "loss": 4.59100866317749,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.011482699774205685,
+ "learning_rate": 0.0023540981982370997,
+ "loss": 4.620088577270508,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.012950294651091099,
+ "learning_rate": 0.0023529981145583483,
+ "loss": 4.578657627105713,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.012197710573673248,
+ "learning_rate": 0.002351897355719652,
+ "loss": 4.620813846588135,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.014839230105280876,
+ "learning_rate": 0.002350795922607872,
+ "loss": 4.615731239318848,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016670791432261467,
+ "learning_rate": 0.002349693816110414,
+ "loss": 4.655855178833008,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.014426100067794323,
+ "learning_rate": 0.0023485910371152247,
+ "loss": 4.660849571228027,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.014040560461580753,
+ "learning_rate": 0.0023474875865107936,
+ "loss": 4.634917736053467,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018269391730427742,
+ "learning_rate": 0.0023463834651861496,
+ "loss": 4.611809253692627,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.013855630531907082,
+ "learning_rate": 0.0023452786740308647,
+ "loss": 4.615225791931152,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.011505477130413055,
+ "learning_rate": 0.0023441732139350487,
+ "loss": 4.62592887878418,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.010166114196181297,
+ "learning_rate": 0.002343067085789351,
+ "loss": 4.6219024658203125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.009253021329641342,
+ "learning_rate": 0.0023419602904849596,
+ "loss": 4.658320426940918,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01027526892721653,
+ "learning_rate": 0.002340852828913598,
+ "loss": 4.589777946472168,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.008855022490024567,
+ "learning_rate": 0.002339744701967529,
+ "loss": 4.588204383850098,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.00965463649481535,
+ "learning_rate": 0.002338635910539551,
+ "loss": 4.648177146911621,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.009073720313608646,
+ "learning_rate": 0.0023375264555229956,
+ "loss": 4.585582733154297,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.010105383582413197,
+ "learning_rate": 0.0023364163378117317,
+ "loss": 4.612174987792969,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.015307556837797165,
+ "learning_rate": 0.002335305558300161,
+ "loss": 4.591984748840332,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016716254875063896,
+ "learning_rate": 0.002334194117883219,
+ "loss": 4.659285545349121,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.013281598687171936,
+ "learning_rate": 0.0023330820174563725,
+ "loss": 4.645829200744629,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.014416513964533806,
+ "learning_rate": 0.0023319692579156204,
+ "loss": 4.584893226623535,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.016326071694493294,
+ "learning_rate": 0.002330855840157494,
+ "loss": 4.633223533630371,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.012851824052631855,
+ "learning_rate": 0.002329741765079053,
+ "loss": 4.580986499786377,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.015444506891071796,
+ "learning_rate": 0.0023286270335778876,
+ "loss": 4.615891456604004,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.01391979493200779,
+ "learning_rate": 0.0023275116465521174,
+ "loss": 4.628403186798096,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.012138686142861843,
+ "learning_rate": 0.0023263956049003895,
+ "loss": 4.6259942054748535,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.013741672039031982,
+ "learning_rate": 0.0023252789095218785,
+ "loss": 4.6056742668151855,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01103319600224495,
+ "learning_rate": 0.002324161561316285,
+ "loss": 4.621670246124268,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.010333316400647163,
+ "learning_rate": 0.0023230435611838367,
+ "loss": 4.620484352111816,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.011063606478273869,
+ "learning_rate": 0.0023219249100252864,
+ "loss": 4.64952278137207,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.009984857402741909,
+ "learning_rate": 0.0023208056087419106,
+ "loss": 4.572571754455566,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.009986856952309608,
+ "learning_rate": 0.002319685658235511,
+ "loss": 4.570366382598877,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.010905452072620392,
+ "learning_rate": 0.0023185650594084103,
+ "loss": 4.603614807128906,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.012800061143934727,
+ "learning_rate": 0.002317443813163456,
+ "loss": 4.562705039978027,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.01252608560025692,
+ "learning_rate": 0.002316321920404015,
+ "loss": 4.599016189575195,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.011305592954158783,
+ "learning_rate": 0.002315199382033976,
+ "loss": 4.644617557525635,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.010594413615763187,
+ "learning_rate": 0.002314076198957749,
+ "loss": 4.569808006286621,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.009859138168394566,
+ "learning_rate": 0.002312952372080261,
+ "loss": 4.5890398025512695,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.010982917621731758,
+ "learning_rate": 0.00231182790230696,
+ "loss": 4.606348991394043,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.010252755135297775,
+ "learning_rate": 0.00231070279054381,
+ "loss": 4.58942174911499,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01110257301479578,
+ "learning_rate": 0.0023095770376972947,
+ "loss": 4.599823474884033,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01003815233707428,
+ "learning_rate": 0.0023084506446744103,
+ "loss": 4.569343090057373,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.010517258197069168,
+ "learning_rate": 0.0023073236123826745,
+ "loss": 4.646301746368408,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.804611682891846,
+ "eval_runtime": 51.6279,
+ "eval_samples_per_second": 47.3,
+ "eval_steps_per_second": 1.491,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.011439066380262375,
+ "learning_rate": 0.0023061959417301145,
+ "loss": 4.563335418701172,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.01073157787322998,
+ "learning_rate": 0.002305067633625275,
+ "loss": 4.540966987609863,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.011391228064894676,
+ "learning_rate": 0.0023039386889772136,
+ "loss": 4.555583953857422,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.012814595364034176,
+ "learning_rate": 0.0023028091086955,
+ "loss": 4.517425537109375,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.012994988821446896,
+ "learning_rate": 0.002301678893690218,
+ "loss": 4.538477897644043,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.01460991520434618,
+ "learning_rate": 0.00230054804487196,
+ "loss": 4.571145057678223,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.013036075979471207,
+ "learning_rate": 0.0022994165631518306,
+ "loss": 4.519233703613281,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.011917057447135448,
+ "learning_rate": 0.0022982844494414444,
+ "loss": 4.5071587562561035,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.013928866013884544,
+ "learning_rate": 0.0022971517046529257,
+ "loss": 4.531422138214111,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01407038513571024,
+ "learning_rate": 0.0022960183296989053,
+ "loss": 4.511205196380615,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.019396567717194557,
+ "learning_rate": 0.0022948843254925245,
+ "loss": 4.530007362365723,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.014425835572183132,
+ "learning_rate": 0.002293749692947428,
+ "loss": 4.530138969421387,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.01600247249007225,
+ "learning_rate": 0.00229261443297777,
+ "loss": 4.5643391609191895,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.018238037824630737,
+ "learning_rate": 0.0022914785464982095,
+ "loss": 4.540779113769531,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.015427535399794579,
+ "learning_rate": 0.0022903420344239085,
+ "loss": 4.506287097930908,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.016816625371575356,
+ "learning_rate": 0.0022892048976705347,
+ "loss": 4.537919521331787,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.01999717950820923,
+ "learning_rate": 0.002288067137154259,
+ "loss": 4.558426380157471,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.020884918048977852,
+ "learning_rate": 0.002286928753791753,
+ "loss": 4.592915058135986,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.01620141603052616,
+ "learning_rate": 0.002285789748500194,
+ "loss": 4.553356170654297,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.018495604395866394,
+ "learning_rate": 0.0022846501221972567,
+ "loss": 4.573139190673828,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.0156320258975029,
+ "learning_rate": 0.0022835098758011174,
+ "loss": 4.585921764373779,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.014515317976474762,
+ "learning_rate": 0.002282369010230452,
+ "loss": 4.5532002449035645,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.014826556667685509,
+ "learning_rate": 0.0022812275264044355,
+ "loss": 4.560998916625977,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.013240831904113293,
+ "learning_rate": 0.00228008542524274,
+ "loss": 4.531787872314453,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.013814719393849373,
+ "learning_rate": 0.0022789427076655372,
+ "loss": 4.611872673034668,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.012246561236679554,
+ "learning_rate": 0.0022777993745934934,
+ "loss": 4.5565385818481445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01206218171864748,
+ "learning_rate": 0.002276655426947771,
+ "loss": 4.570852279663086,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01007329672574997,
+ "learning_rate": 0.0022755108656500275,
+ "loss": 4.553400039672852,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.010361720807850361,
+ "learning_rate": 0.0022743656916224173,
+ "loss": 4.527509689331055,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.009456256404519081,
+ "learning_rate": 0.0022732199057875846,
+ "loss": 4.520236015319824,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.009310049936175346,
+ "learning_rate": 0.002272073509068669,
+ "loss": 4.518343925476074,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.00927449855953455,
+ "learning_rate": 0.0022709265023893015,
+ "loss": 4.559690475463867,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.008658594451844692,
+ "learning_rate": 0.002269778886673605,
+ "loss": 4.513876914978027,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.008928220719099045,
+ "learning_rate": 0.0022686306628461918,
+ "loss": 4.545050621032715,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.008046358823776245,
+ "learning_rate": 0.002267481831832166,
+ "loss": 4.54866361618042,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.008318924345076084,
+ "learning_rate": 0.00226633239455712,
+ "loss": 4.546968936920166,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.008808869868516922,
+ "learning_rate": 0.0022651823519471343,
+ "loss": 4.50877571105957,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.008767383173108101,
+ "learning_rate": 0.0022640317049287774,
+ "loss": 4.57666540145874,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.009351354092359543,
+ "learning_rate": 0.0022628804544291047,
+ "loss": 4.511702537536621,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.010137057863175869,
+ "learning_rate": 0.002261728601375659,
+ "loss": 4.539676189422607,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.00882046204060316,
+ "learning_rate": 0.002260576146696466,
+ "loss": 4.533196449279785,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.009354782290756702,
+ "learning_rate": 0.0022594230913200388,
+ "loss": 4.515488624572754,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.009241201914846897,
+ "learning_rate": 0.002258269436175372,
+ "loss": 4.542706489562988,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.009252829477190971,
+ "learning_rate": 0.002257115182191947,
+ "loss": 4.592660903930664,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.011243131943047047,
+ "learning_rate": 0.0022559603302997225,
+ "loss": 4.527804851531982,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.011032823473215103,
+ "learning_rate": 0.0022548048814291437,
+ "loss": 4.540122032165527,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.010495275259017944,
+ "learning_rate": 0.0022536488365111347,
+ "loss": 4.532835006713867,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.009992090053856373,
+ "learning_rate": 0.0022524921964770994,
+ "loss": 4.521217346191406,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01141198631376028,
+ "learning_rate": 0.0022513349622589225,
+ "loss": 4.519709587097168,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.013087889179587364,
+ "learning_rate": 0.002250177134788966,
+ "loss": 4.547289848327637,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.01328477542847395,
+ "learning_rate": 0.0022490187150000706,
+ "loss": 4.527865409851074,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01033794041723013,
+ "learning_rate": 0.002247859703825555,
+ "loss": 4.54740047454834,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.010522300377488136,
+ "learning_rate": 0.002246700102199213,
+ "loss": 4.538840293884277,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.012151090428233147,
+ "learning_rate": 0.0022455399110553145,
+ "loss": 4.536865234375,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.010553102008998394,
+ "learning_rate": 0.0022443791313286054,
+ "loss": 4.5348334312438965,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.011433709412813187,
+ "learning_rate": 0.0022432177639543035,
+ "loss": 4.510526657104492,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01189691387116909,
+ "learning_rate": 0.002242055809868103,
+ "loss": 4.530807018280029,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.010031254962086678,
+ "learning_rate": 0.0022408932700061683,
+ "loss": 4.496850967407227,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.010129331611096859,
+ "learning_rate": 0.002239730145305138,
+ "loss": 4.551287651062012,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.010430586524307728,
+ "learning_rate": 0.0022385664367021185,
+ "loss": 4.542672157287598,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.011774476617574692,
+ "learning_rate": 0.002237402145134691,
+ "loss": 4.537034034729004,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.01184659544378519,
+ "learning_rate": 0.002236237271540903,
+ "loss": 4.521193504333496,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.011327247135341167,
+ "learning_rate": 0.0022350718168592724,
+ "loss": 4.49426794052124,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.010767918080091476,
+ "learning_rate": 0.0022339057820287855,
+ "loss": 4.531225204467773,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.011195424944162369,
+ "learning_rate": 0.002232739167988894,
+ "loss": 4.492671012878418,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.010394908487796783,
+ "learning_rate": 0.0022315719756795197,
+ "loss": 4.523616790771484,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.010277998633682728,
+ "learning_rate": 0.002230404206041047,
+ "loss": 4.487832069396973,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.010955783538520336,
+ "learning_rate": 0.0022292358600143275,
+ "loss": 4.526865005493164,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.010611076839268208,
+ "learning_rate": 0.002228066938540676,
+ "loss": 4.498682022094727,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.010008828714489937,
+ "learning_rate": 0.0022268974425618724,
+ "loss": 4.524019241333008,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.008973540738224983,
+ "learning_rate": 0.002225727373020157,
+ "loss": 4.512996673583984,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.011754963546991348,
+ "learning_rate": 0.002224556730858235,
+ "loss": 4.496376991271973,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.744190216064453,
+ "eval_runtime": 51.6572,
+ "eval_samples_per_second": 47.273,
+ "eval_steps_per_second": 1.491,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.012094628997147083,
+ "learning_rate": 0.0022233855170192703,
+ "loss": 4.443447113037109,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.014347990974783897,
+ "learning_rate": 0.00222221373244689,
+ "loss": 4.485690116882324,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.021286984905600548,
+ "learning_rate": 0.0022210413780851786,
+ "loss": 4.487505912780762,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.020273910835385323,
+ "learning_rate": 0.0022198684548786827,
+ "loss": 4.473119735717773,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.017642971128225327,
+ "learning_rate": 0.002218694963772403,
+ "loss": 4.440615653991699,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.02075147069990635,
+ "learning_rate": 0.0022175209057118017,
+ "loss": 4.527026176452637,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016479508951306343,
+ "learning_rate": 0.002216346281642796,
+ "loss": 4.508321762084961,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.017978593707084656,
+ "learning_rate": 0.002215171092511759,
+ "loss": 4.500985145568848,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.01739897020161152,
+ "learning_rate": 0.002213995339265519,
+ "loss": 4.474225044250488,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.015556776896119118,
+ "learning_rate": 0.0022128190228513597,
+ "loss": 4.444849014282227,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.016166578978300095,
+ "learning_rate": 0.002211642144217018,
+ "loss": 4.5026726722717285,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.01534261740744114,
+ "learning_rate": 0.0022104647043106832,
+ "loss": 4.484527587890625,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.013517106883227825,
+ "learning_rate": 0.002209286704080998,
+ "loss": 4.453907012939453,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.012562787160277367,
+ "learning_rate": 0.0022081081444770556,
+ "loss": 4.471528053283691,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.009947682730853558,
+ "learning_rate": 0.0022069290264483997,
+ "loss": 4.504153728485107,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01169571466743946,
+ "learning_rate": 0.002205749350945026,
+ "loss": 4.447345733642578,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.011394352652132511,
+ "learning_rate": 0.0022045691189173755,
+ "loss": 4.447491645812988,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.011168593540787697,
+ "learning_rate": 0.0022033883313163417,
+ "loss": 4.441639423370361,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.009503341279923916,
+ "learning_rate": 0.002202206989093263,
+ "loss": 4.461601257324219,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.0112199941650033,
+ "learning_rate": 0.002201025093199926,
+ "loss": 4.444095134735107,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.010505333542823792,
+ "learning_rate": 0.002199842644588562,
+ "loss": 4.460987091064453,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.009494813159108162,
+ "learning_rate": 0.0021986596442118496,
+ "loss": 4.460158348083496,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.01063452661037445,
+ "learning_rate": 0.00219747609302291,
+ "loss": 4.489625453948975,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.012194888666272163,
+ "learning_rate": 0.0021962919919753087,
+ "loss": 4.441605567932129,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.010613647289574146,
+ "learning_rate": 0.0021951073420230557,
+ "loss": 4.475821018218994,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.009643700905144215,
+ "learning_rate": 0.002193922144120601,
+ "loss": 4.467228889465332,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.00912613607943058,
+ "learning_rate": 0.0021927363992228373,
+ "loss": 4.455435752868652,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.009371735155582428,
+ "learning_rate": 0.0021915501082850988,
+ "loss": 4.479915142059326,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.008617875166237354,
+ "learning_rate": 0.002190363272263158,
+ "loss": 4.415642738342285,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.009571383707225323,
+ "learning_rate": 0.002189175892113227,
+ "loss": 4.439854621887207,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.00782635249197483,
+ "learning_rate": 0.002187987968791957,
+ "loss": 4.484311103820801,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.008408057503402233,
+ "learning_rate": 0.0021867995032564358,
+ "loss": 4.46577262878418,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.009641910903155804,
+ "learning_rate": 0.00218561049646419,
+ "loss": 4.437408447265625,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.008498182520270348,
+ "learning_rate": 0.0021844209493731784,
+ "loss": 4.442599296569824,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.009289798326790333,
+ "learning_rate": 0.0021832308629418,
+ "loss": 4.517641067504883,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.009360183961689472,
+ "learning_rate": 0.0021820402381288848,
+ "loss": 4.474143981933594,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.008764738216996193,
+ "learning_rate": 0.0021808490758936983,
+ "loss": 4.4376702308654785,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.009673701599240303,
+ "learning_rate": 0.0021796573771959375,
+ "loss": 4.435220718383789,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.010852573439478874,
+ "learning_rate": 0.0021784651429957334,
+ "loss": 4.465187072753906,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.011292314156889915,
+ "learning_rate": 0.0021772723742536464,
+ "loss": 4.458176136016846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01071334257721901,
+ "learning_rate": 0.0021760790719306702,
+ "loss": 4.425498962402344,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.01505239401012659,
+ "learning_rate": 0.0021748852369882265,
+ "loss": 4.463129043579102,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.011774689890444279,
+ "learning_rate": 0.0021736908703881663,
+ "loss": 4.462577819824219,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.011241118423640728,
+ "learning_rate": 0.002172495973092769,
+ "loss": 4.452583312988281,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.01072910986840725,
+ "learning_rate": 0.002171300546064742,
+ "loss": 4.466248035430908,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01021308358758688,
+ "learning_rate": 0.002170104590267219,
+ "loss": 4.485857963562012,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.012630444951355457,
+ "learning_rate": 0.00216890810666376,
+ "loss": 4.442112922668457,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.013602569699287415,
+ "learning_rate": 0.0021677110962183515,
+ "loss": 4.45171594619751,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.013808641582727432,
+ "learning_rate": 0.0021665135598954014,
+ "loss": 4.468818187713623,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.011510699987411499,
+ "learning_rate": 0.002165315498659744,
+ "loss": 4.439584255218506,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.010502424091100693,
+ "learning_rate": 0.0021641169134766353,
+ "loss": 4.478377342224121,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.009953407570719719,
+ "learning_rate": 0.0021629178053117542,
+ "loss": 4.474404335021973,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.011158930137753487,
+ "learning_rate": 0.0021617181751312,
+ "loss": 4.4549560546875,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.010431940667331219,
+ "learning_rate": 0.002160518023901493,
+ "loss": 4.444911003112793,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.009100335650146008,
+ "learning_rate": 0.0021593173525895734,
+ "loss": 4.428071975708008,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.009362847544252872,
+ "learning_rate": 0.0021581161621628013,
+ "loss": 4.462827682495117,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.009893129579722881,
+ "learning_rate": 0.0021569144535889535,
+ "loss": 4.426912307739258,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.010524468496441841,
+ "learning_rate": 0.0021557122278362244,
+ "loss": 4.435606002807617,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.010621445253491402,
+ "learning_rate": 0.002154509485873226,
+ "loss": 4.454901695251465,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.010836896486580372,
+ "learning_rate": 0.0021533062286689855,
+ "loss": 4.4441728591918945,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.011403030715882778,
+ "learning_rate": 0.0021521024571929457,
+ "loss": 4.483884811401367,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.012522426433861256,
+ "learning_rate": 0.0021508981724149643,
+ "loss": 4.464006423950195,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.014799179509282112,
+ "learning_rate": 0.0021496933753053105,
+ "loss": 4.460402011871338,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.012453878298401833,
+ "learning_rate": 0.0021484880668346685,
+ "loss": 4.456899642944336,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.012270379811525345,
+ "learning_rate": 0.002147282247974133,
+ "loss": 4.477325439453125,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01094560045748949,
+ "learning_rate": 0.0021460759196952105,
+ "loss": 4.422362327575684,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.009995176456868649,
+ "learning_rate": 0.0021448690829698183,
+ "loss": 4.405127048492432,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.014935536310076714,
+ "learning_rate": 0.002143661738770283,
+ "loss": 4.445840358734131,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.011492030695080757,
+ "learning_rate": 0.0021424538880693393,
+ "loss": 4.459555625915527,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.008793608285486698,
+ "learning_rate": 0.002141245531840132,
+ "loss": 4.436631202697754,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.008894025348126888,
+ "learning_rate": 0.0021400366710562103,
+ "loss": 4.473924160003662,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.010324863716959953,
+ "learning_rate": 0.0021388273066915327,
+ "loss": 4.475680351257324,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.688324451446533,
+ "eval_runtime": 54.4515,
+ "eval_samples_per_second": 44.847,
+ "eval_steps_per_second": 1.414,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.009235628880560398,
+ "learning_rate": 0.0021376174397204615,
+ "loss": 4.370064735412598,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.008818200789391994,
+ "learning_rate": 0.0021364070711177643,
+ "loss": 4.371678352355957,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.010333757847547531,
+ "learning_rate": 0.0021351962018586147,
+ "loss": 4.3896379470825195,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.011938809417188168,
+ "learning_rate": 0.002133984832918587,
+ "loss": 4.361343860626221,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.010634614154696465,
+ "learning_rate": 0.00213277296527366,
+ "loss": 4.362449645996094,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.008545687422156334,
+ "learning_rate": 0.0021315605999002126,
+ "loss": 4.3685503005981445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.009159795008599758,
+ "learning_rate": 0.002130347737775027,
+ "loss": 4.379969596862793,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.010464153252542019,
+ "learning_rate": 0.0021291343798752835,
+ "loss": 4.376113414764404,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.011010527610778809,
+ "learning_rate": 0.0021279205271785622,
+ "loss": 4.390711784362793,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.010350867174565792,
+ "learning_rate": 0.002126706180662843,
+ "loss": 4.399683952331543,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.009708237834274769,
+ "learning_rate": 0.002125491341306503,
+ "loss": 4.363435745239258,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.010087115690112114,
+ "learning_rate": 0.002124276010088316,
+ "loss": 4.366623878479004,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.009712349623441696,
+ "learning_rate": 0.002123060187987453,
+ "loss": 4.360814571380615,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.010174986906349659,
+ "learning_rate": 0.0021218438759834785,
+ "loss": 4.347475051879883,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.010195785202085972,
+ "learning_rate": 0.0021206270750563543,
+ "loss": 4.399251461029053,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.010786959901452065,
+ "learning_rate": 0.0021194097861864354,
+ "loss": 4.397933006286621,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.011476189829409122,
+ "learning_rate": 0.002118192010354468,
+ "loss": 4.3748369216918945,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.011571013368666172,
+ "learning_rate": 0.0021169737485415927,
+ "loss": 4.388349533081055,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.011140935122966766,
+ "learning_rate": 0.002115755001729341,
+ "loss": 4.398018836975098,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.013216734863817692,
+ "learning_rate": 0.0021145357708996365,
+ "loss": 4.406947135925293,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.013385126367211342,
+ "learning_rate": 0.0021133160570347893,
+ "loss": 4.41457462310791,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.011784231290221214,
+ "learning_rate": 0.002112095861117502,
+ "loss": 4.403835296630859,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.009813498705625534,
+ "learning_rate": 0.0021108751841308645,
+ "loss": 4.3865647315979,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.010297936387360096,
+ "learning_rate": 0.0021096540270583533,
+ "loss": 4.395583152770996,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.011965147219598293,
+ "learning_rate": 0.002108432390883834,
+ "loss": 4.385443687438965,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.012212193571031094,
+ "learning_rate": 0.0021072102765915556,
+ "loss": 4.422321319580078,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.009840975515544415,
+ "learning_rate": 0.0021059876851661544,
+ "loss": 4.3568902015686035,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.013039806857705116,
+ "learning_rate": 0.002104764617592649,
+ "loss": 4.314867973327637,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.015091902576386929,
+ "learning_rate": 0.0021035410748564444,
+ "loss": 4.392338752746582,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.015448705293238163,
+ "learning_rate": 0.0021023170579433257,
+ "loss": 4.379609107971191,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.018393104895949364,
+ "learning_rate": 0.0021010925678394622,
+ "loss": 4.418291091918945,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.01866372860968113,
+ "learning_rate": 0.002099867605531402,
+ "loss": 4.4006452560424805,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.01819106750190258,
+ "learning_rate": 0.002098642172006077,
+ "loss": 4.433975696563721,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.019774602726101875,
+ "learning_rate": 0.0020974162682507952,
+ "loss": 4.417821884155273,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.016885802149772644,
+ "learning_rate": 0.0020961898952532457,
+ "loss": 4.397073268890381,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.015750881284475327,
+ "learning_rate": 0.0020949630540014956,
+ "loss": 4.4356608390808105,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.015696514397859573,
+ "learning_rate": 0.002093735745483988,
+ "loss": 4.422449588775635,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.017195021733641624,
+ "learning_rate": 0.0020925079706895435,
+ "loss": 4.443209648132324,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.015864048153162003,
+ "learning_rate": 0.0020912797306073577,
+ "loss": 4.4340691566467285,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.009857800789177418,
+ "learning_rate": 0.0020900510262270022,
+ "loss": 4.37935209274292,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.013739476911723614,
+ "learning_rate": 0.0020888218585384216,
+ "loss": 4.412148475646973,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.011667690239846706,
+ "learning_rate": 0.0020875922285319344,
+ "loss": 4.391700744628906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.012323460541665554,
+ "learning_rate": 0.0020863621371982302,
+ "loss": 4.423880577087402,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.010353626683354378,
+ "learning_rate": 0.0020851315855283724,
+ "loss": 4.427382469177246,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.0101751908659935,
+ "learning_rate": 0.002083900574513794,
+ "loss": 4.415317535400391,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.01126549206674099,
+ "learning_rate": 0.002082669105146298,
+ "loss": 4.402498245239258,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.011096738278865814,
+ "learning_rate": 0.0020814371784180575,
+ "loss": 4.433567047119141,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.010944445617496967,
+ "learning_rate": 0.0020802047953216137,
+ "loss": 4.406990051269531,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.008965318091213703,
+ "learning_rate": 0.0020789719568498747,
+ "loss": 4.3721771240234375,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.00828113779425621,
+ "learning_rate": 0.002077738663996117,
+ "loss": 4.389828205108643,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.008599454537034035,
+ "learning_rate": 0.002076504917753982,
+ "loss": 4.411396026611328,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.007713889703154564,
+ "learning_rate": 0.0020752707191174762,
+ "loss": 4.355343341827393,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.00912355538457632,
+ "learning_rate": 0.0020740360690809723,
+ "loss": 4.423413276672363,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.008269775658845901,
+ "learning_rate": 0.0020728009686392045,
+ "loss": 4.390920639038086,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.007461601868271828,
+ "learning_rate": 0.002071565418787272,
+ "loss": 4.365501403808594,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.007951333187520504,
+ "learning_rate": 0.0020703294205206336,
+ "loss": 4.378884315490723,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.007816503755748272,
+ "learning_rate": 0.002069092974835112,
+ "loss": 4.4241743087768555,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.00833756010979414,
+ "learning_rate": 0.0020678560827268886,
+ "loss": 4.391102313995361,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.008241684176027775,
+ "learning_rate": 0.0020666187451925055,
+ "loss": 4.374970436096191,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.008245996199548244,
+ "learning_rate": 0.002065380963228862,
+ "loss": 4.380222320556641,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.007378706708550453,
+ "learning_rate": 0.0020641427378332183,
+ "loss": 4.391451358795166,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.007594958413392305,
+ "learning_rate": 0.002062904070003189,
+ "loss": 4.372774124145508,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.0076752156019210815,
+ "learning_rate": 0.002061664960736747,
+ "loss": 4.384668827056885,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.008727089501917362,
+ "learning_rate": 0.00206042541103222,
+ "loss": 4.363401412963867,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.007616396062076092,
+ "learning_rate": 0.002059185421888291,
+ "loss": 4.413188934326172,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.008922954089939594,
+ "learning_rate": 0.0020579449943039967,
+ "loss": 4.383417129516602,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.009063739329576492,
+ "learning_rate": 0.002056704129278727,
+ "loss": 4.385283470153809,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.008912574499845505,
+ "learning_rate": 0.002055462827812225,
+ "loss": 4.378478050231934,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.008317525498569012,
+ "learning_rate": 0.0020542210909045844,
+ "loss": 4.367859840393066,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.009347227402031422,
+ "learning_rate": 0.0020529789195562506,
+ "loss": 4.38238000869751,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.011030065827071667,
+ "learning_rate": 0.0020517363147680183,
+ "loss": 4.372624397277832,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.011519700288772583,
+ "learning_rate": 0.002050493277541032,
+ "loss": 4.3906121253967285,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.641542434692383,
+ "eval_runtime": 51.6342,
+ "eval_samples_per_second": 47.294,
+ "eval_steps_per_second": 1.491,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.009694516658782959,
+ "learning_rate": 0.0020492498088767836,
+ "loss": 4.32808256149292,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.0097952950745821,
+ "learning_rate": 0.002048005909777114,
+ "loss": 4.280096054077148,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.010393359698355198,
+ "learning_rate": 0.0020467615812442106,
+ "loss": 4.317824363708496,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.00977372657507658,
+ "learning_rate": 0.0020455168242806054,
+ "loss": 4.3445329666137695,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.008568719029426575,
+ "learning_rate": 0.0020442716398891775,
+ "loss": 4.318026542663574,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.009732510894536972,
+ "learning_rate": 0.0020430260290731487,
+ "loss": 4.330609321594238,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.010429969057440758,
+ "learning_rate": 0.0020417799928360864,
+ "loss": 4.316794395446777,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.009292383678257465,
+ "learning_rate": 0.002040533532181898,
+ "loss": 4.317420959472656,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.011222455650568008,
+ "learning_rate": 0.0020392866481148356,
+ "loss": 4.292610168457031,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.013591725379228592,
+ "learning_rate": 0.0020380393416394894,
+ "loss": 4.325689315795898,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.013572026044130325,
+ "learning_rate": 0.002036791613760794,
+ "loss": 4.334960460662842,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.012986521236598492,
+ "learning_rate": 0.0020355434654840197,
+ "loss": 4.33829402923584,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.013321424834430218,
+ "learning_rate": 0.0020342948978147777,
+ "loss": 4.2926926612854,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.012307215481996536,
+ "learning_rate": 0.0020330459117590153,
+ "loss": 4.32757568359375,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.012909360229969025,
+ "learning_rate": 0.0020317965083230197,
+ "loss": 4.362889289855957,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017746610566973686,
+ "learning_rate": 0.0020305466885134115,
+ "loss": 4.3426313400268555,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.01696564070880413,
+ "learning_rate": 0.002029296453337148,
+ "loss": 4.3298516273498535,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.013348255306482315,
+ "learning_rate": 0.002028045803801522,
+ "loss": 4.360220432281494,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.013225145637989044,
+ "learning_rate": 0.002026794740914158,
+ "loss": 4.346250057220459,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.01569020375609398,
+ "learning_rate": 0.002025543265683016,
+ "loss": 4.308025360107422,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.018612321466207504,
+ "learning_rate": 0.002024291379116385,
+ "loss": 4.330190181732178,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.01619943603873253,
+ "learning_rate": 0.00202303908222289,
+ "loss": 4.339552402496338,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.015385213308036327,
+ "learning_rate": 0.0020217863760114826,
+ "loss": 4.3412184715271,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.018517112359404564,
+ "learning_rate": 0.002020533261491446,
+ "loss": 4.3289313316345215,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.014868509024381638,
+ "learning_rate": 0.0020192797396723903,
+ "loss": 4.359846115112305,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.014240248128771782,
+ "learning_rate": 0.0020180258115642578,
+ "loss": 4.369707107543945,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.018113158643245697,
+ "learning_rate": 0.0020167714781773143,
+ "loss": 4.362668514251709,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.01802460104227066,
+ "learning_rate": 0.002015516740522154,
+ "loss": 4.344742298126221,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.01414549071341753,
+ "learning_rate": 0.0020142615996096964,
+ "loss": 4.358080863952637,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.012157700955867767,
+ "learning_rate": 0.002013006056451186,
+ "loss": 4.347763538360596,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.011268534697592258,
+ "learning_rate": 0.0020117501120581905,
+ "loss": 4.335859298706055,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.010395628400146961,
+ "learning_rate": 0.002010493767442602,
+ "loss": 4.341609001159668,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.012147950939834118,
+ "learning_rate": 0.002009237023616635,
+ "loss": 4.328502655029297,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.00945813488215208,
+ "learning_rate": 0.0020079798815928252,
+ "loss": 4.311186790466309,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.009489500895142555,
+ "learning_rate": 0.0020067223423840294,
+ "loss": 4.371470928192139,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.009360029362142086,
+ "learning_rate": 0.0020054644070034234,
+ "loss": 4.33941125869751,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.010032779537141323,
+ "learning_rate": 0.0020042060764645037,
+ "loss": 4.365797519683838,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.009367046877741814,
+ "learning_rate": 0.002002947351781084,
+ "loss": 4.29443359375,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.008369733579456806,
+ "learning_rate": 0.0020016882339672967,
+ "loss": 4.294233322143555,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.010394785553216934,
+ "learning_rate": 0.0020004287240375893,
+ "loss": 4.352355003356934,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01068816613405943,
+ "learning_rate": 0.001999168823006728,
+ "loss": 4.342516899108887,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.009975973516702652,
+ "learning_rate": 0.0019979085318897905,
+ "loss": 4.346763610839844,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.009618151932954788,
+ "learning_rate": 0.001996647851702171,
+ "loss": 4.329115867614746,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.008399694226682186,
+ "learning_rate": 0.001995386783459577,
+ "loss": 4.28736686706543,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.008235268294811249,
+ "learning_rate": 0.001994125328178029,
+ "loss": 4.339872360229492,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.00836547277867794,
+ "learning_rate": 0.0019928634868738577,
+ "loss": 4.328156471252441,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.009619626216590405,
+ "learning_rate": 0.001991601260563706,
+ "loss": 4.337394714355469,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.009814038872718811,
+ "learning_rate": 0.0019903386502645275,
+ "loss": 4.325926780700684,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.009588954970240593,
+ "learning_rate": 0.001989075656993584,
+ "loss": 4.286477088928223,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.010471943765878677,
+ "learning_rate": 0.001987812281768447,
+ "loss": 4.322466850280762,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.010239035822451115,
+ "learning_rate": 0.001986548525606994,
+ "loss": 4.3487653732299805,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.008935568854212761,
+ "learning_rate": 0.001985284389527411,
+ "loss": 4.303601264953613,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.008624759502708912,
+ "learning_rate": 0.00198401987454819,
+ "loss": 4.34036922454834,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.009296547621488571,
+ "learning_rate": 0.0019827549816881277,
+ "loss": 4.273357391357422,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.009071138687431812,
+ "learning_rate": 0.0019814897119663248,
+ "loss": 4.321051597595215,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.008341389708220959,
+ "learning_rate": 0.0019802240664021876,
+ "loss": 4.343242645263672,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.007881715893745422,
+ "learning_rate": 0.0019789580460154227,
+ "loss": 4.32067346572876,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.008750553242862225,
+ "learning_rate": 0.0019776916518260404,
+ "loss": 4.330263614654541,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.008591460064053535,
+ "learning_rate": 0.001976424884854351,
+ "loss": 4.311957359313965,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.008084543980658054,
+ "learning_rate": 0.001975157746120966,
+ "loss": 4.331319808959961,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.008036594837903976,
+ "learning_rate": 0.001973890236646797,
+ "loss": 4.382656097412109,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.00805772002786398,
+ "learning_rate": 0.0019726223574530527,
+ "loss": 4.297247886657715,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.009019420482218266,
+ "learning_rate": 0.0019713541095612405,
+ "loss": 4.326526165008545,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.009953883476555347,
+ "learning_rate": 0.0019700854939931653,
+ "loss": 4.327356338500977,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.009340492077171803,
+ "learning_rate": 0.0019688165117709273,
+ "loss": 4.343993186950684,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.010395721532404423,
+ "learning_rate": 0.0019675471639169226,
+ "loss": 4.284934997558594,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.009946230798959732,
+ "learning_rate": 0.0019662774514538415,
+ "loss": 4.288936614990234,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.010981100611388683,
+ "learning_rate": 0.0019650073754046695,
+ "loss": 4.339935779571533,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.009688783437013626,
+ "learning_rate": 0.0019637369367926827,
+ "loss": 4.381425857543945,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.009442255832254887,
+ "learning_rate": 0.001962466136641452,
+ "loss": 4.304185390472412,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.009371291846036911,
+ "learning_rate": 0.0019611949759748364,
+ "loss": 4.306120872497559,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.009544814005494118,
+ "learning_rate": 0.001959923455816988,
+ "loss": 4.326251983642578,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.602745532989502,
+ "eval_runtime": 53.5041,
+ "eval_samples_per_second": 45.641,
+ "eval_steps_per_second": 1.439,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.01010364294052124,
+ "learning_rate": 0.0019586515771923474,
+ "loss": 4.24491024017334,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.012722729705274105,
+ "learning_rate": 0.0019573793411256446,
+ "loss": 4.257739543914795,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.01307204831391573,
+ "learning_rate": 0.0019561067486418967,
+ "loss": 4.291684150695801,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.011215966194868088,
+ "learning_rate": 0.0019548338007664098,
+ "loss": 4.237947463989258,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.012943019159138203,
+ "learning_rate": 0.0019535604985247727,
+ "loss": 4.242859363555908,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.01337282732129097,
+ "learning_rate": 0.0019522868429428636,
+ "loss": 4.238934516906738,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.013697991147637367,
+ "learning_rate": 0.0019510128350468438,
+ "loss": 4.285760402679443,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.012182159349322319,
+ "learning_rate": 0.001949738475863158,
+ "loss": 4.290044784545898,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.012620339170098305,
+ "learning_rate": 0.0019484637664185346,
+ "loss": 4.267187118530273,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.011284958571195602,
+ "learning_rate": 0.0019471887077399837,
+ "loss": 4.26662540435791,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.01161450706422329,
+ "learning_rate": 0.001945913300854797,
+ "loss": 4.283192157745361,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.010022968985140324,
+ "learning_rate": 0.001944637546790548,
+ "loss": 4.295843124389648,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.010341878980398178,
+ "learning_rate": 0.001943361446575087,
+ "loss": 4.295636177062988,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.011549185030162334,
+ "learning_rate": 0.0019420850012365457,
+ "loss": 4.288768768310547,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.012518810108304024,
+ "learning_rate": 0.0019408082118033331,
+ "loss": 4.229338645935059,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.012035328894853592,
+ "learning_rate": 0.0019395310793041353,
+ "loss": 4.258639335632324,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.011193782091140747,
+ "learning_rate": 0.0019382536047679146,
+ "loss": 4.239264488220215,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.009930136613547802,
+ "learning_rate": 0.0019369757892239096,
+ "loss": 4.252286434173584,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.00953229982405901,
+ "learning_rate": 0.0019356976337016327,
+ "loss": 4.293435096740723,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.010618340224027634,
+ "learning_rate": 0.0019344191392308716,
+ "loss": 4.235588550567627,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.008806651458144188,
+ "learning_rate": 0.0019331403068416856,
+ "loss": 4.26716423034668,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.008934796787798405,
+ "learning_rate": 0.0019318611375644065,
+ "loss": 4.258462905883789,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.009283283725380898,
+ "learning_rate": 0.0019305816324296386,
+ "loss": 4.257868766784668,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.009224812500178814,
+ "learning_rate": 0.0019293017924682555,
+ "loss": 4.258486747741699,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.008957969956099987,
+ "learning_rate": 0.0019280216187114016,
+ "loss": 4.309563636779785,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.010284051299095154,
+ "learning_rate": 0.0019267411121904894,
+ "loss": 4.275271415710449,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.009167762473225594,
+ "learning_rate": 0.0019254602739372001,
+ "loss": 4.238452911376953,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.009411636739969254,
+ "learning_rate": 0.0019241791049834816,
+ "loss": 4.2735595703125,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.009044171310961246,
+ "learning_rate": 0.0019228976063615497,
+ "loss": 4.27017879486084,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.008840068243443966,
+ "learning_rate": 0.0019216157791038834,
+ "loss": 4.28773307800293,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.009512359276413918,
+ "learning_rate": 0.001920333624243228,
+ "loss": 4.266008377075195,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.009946667589247227,
+ "learning_rate": 0.001919051142812593,
+ "loss": 4.279062271118164,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.00899894442409277,
+ "learning_rate": 0.0019177683358452505,
+ "loss": 4.277679443359375,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.00909668579697609,
+ "learning_rate": 0.0019164852043747338,
+ "loss": 4.311397552490234,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.009787327609956264,
+ "learning_rate": 0.0019152017494348408,
+ "loss": 4.30875825881958,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.008983899839222431,
+ "learning_rate": 0.001913917972059626,
+ "loss": 4.259649276733398,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.008411582559347153,
+ "learning_rate": 0.0019126338732834071,
+ "loss": 4.259811878204346,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.009162918664515018,
+ "learning_rate": 0.0019113494541407579,
+ "loss": 4.271515846252441,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.011304561980068684,
+ "learning_rate": 0.0019100647156665125,
+ "loss": 4.269217491149902,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.012483394704759121,
+ "learning_rate": 0.0019087796588957617,
+ "loss": 4.2949299812316895,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.012091380544006824,
+ "learning_rate": 0.0019074942848638522,
+ "loss": 4.289752006530762,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.010011567734181881,
+ "learning_rate": 0.0019062085946063869,
+ "loss": 4.294095993041992,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.008766217157244682,
+ "learning_rate": 0.0019049225891592226,
+ "loss": 4.256425857543945,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.012862488627433777,
+ "learning_rate": 0.001903636269558471,
+ "loss": 4.287847995758057,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.013917063362896442,
+ "learning_rate": 0.0019023496368404972,
+ "loss": 4.282784938812256,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.012129023671150208,
+ "learning_rate": 0.0019010626920419173,
+ "loss": 4.2486419677734375,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.010348426178097725,
+ "learning_rate": 0.0018997754361995986,
+ "loss": 4.289836883544922,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.011821693740785122,
+ "learning_rate": 0.001898487870350662,
+ "loss": 4.214344024658203,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.013136724010109901,
+ "learning_rate": 0.0018971999955324742,
+ "loss": 4.24208402633667,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.013173168525099754,
+ "learning_rate": 0.0018959118127826538,
+ "loss": 4.270164489746094,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.011018159799277782,
+ "learning_rate": 0.0018946233231390656,
+ "loss": 4.2571916580200195,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.01018658559769392,
+ "learning_rate": 0.0018933345276398227,
+ "loss": 4.283267498016357,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.009412688203155994,
+ "learning_rate": 0.0018920454273232848,
+ "loss": 4.272198677062988,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.007799267303198576,
+ "learning_rate": 0.0018907560232280557,
+ "loss": 4.260125637054443,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.008747464045882225,
+ "learning_rate": 0.0018894663163929863,
+ "loss": 4.2539215087890625,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.008003434166312218,
+ "learning_rate": 0.0018881763078571689,
+ "loss": 4.245024681091309,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.007593846879899502,
+ "learning_rate": 0.0018868859986599415,
+ "loss": 4.237929344177246,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.008099420927464962,
+ "learning_rate": 0.0018855953898408815,
+ "loss": 4.2789106369018555,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.008270486257970333,
+ "learning_rate": 0.0018843044824398097,
+ "loss": 4.263522148132324,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.009260379709303379,
+ "learning_rate": 0.0018830132774967866,
+ "loss": 4.275835037231445,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.009289492852985859,
+ "learning_rate": 0.0018817217760521132,
+ "loss": 4.313778400421143,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.009493921883404255,
+ "learning_rate": 0.0018804299791463281,
+ "loss": 4.303672790527344,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.009020023047924042,
+ "learning_rate": 0.0018791378878202105,
+ "loss": 4.245473384857178,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.009446556679904461,
+ "learning_rate": 0.0018778455031147724,
+ "loss": 4.267327785491943,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.009499974548816681,
+ "learning_rate": 0.0018765528260712672,
+ "loss": 4.301205635070801,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.008399393409490585,
+ "learning_rate": 0.00187525985773118,
+ "loss": 4.307802677154541,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.009519701823592186,
+ "learning_rate": 0.0018739665991362327,
+ "loss": 4.284796714782715,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.01125232595950365,
+ "learning_rate": 0.0018726730513283804,
+ "loss": 4.285930633544922,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.011608537286520004,
+ "learning_rate": 0.00187137921534981,
+ "loss": 4.257689476013184,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.01064533181488514,
+ "learning_rate": 0.001870085092242944,
+ "loss": 4.249063968658447,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.009854599833488464,
+ "learning_rate": 0.0018687906830504321,
+ "loss": 4.255159378051758,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.01121739111840725,
+ "learning_rate": 0.0018674959888151576,
+ "loss": 4.287777900695801,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.5675368309021,
+ "eval_runtime": 51.8555,
+ "eval_samples_per_second": 47.092,
+ "eval_steps_per_second": 1.485,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.0104464590549469,
+ "learning_rate": 0.0018662010105802314,
+ "loss": 4.206080436706543,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.012800181284546852,
+ "learning_rate": 0.0018649057493889945,
+ "loss": 4.213550567626953,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.013382901437580585,
+ "learning_rate": 0.001863610206285015,
+ "loss": 4.190858840942383,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.014347050338983536,
+ "learning_rate": 0.0018623143823120894,
+ "loss": 4.22884464263916,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.01347152329981327,
+ "learning_rate": 0.001861018278514239,
+ "loss": 4.215579032897949,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.012786377221345901,
+ "learning_rate": 0.0018597218959357112,
+ "loss": 4.174999237060547,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.014137202873826027,
+ "learning_rate": 0.001858425235620979,
+ "loss": 4.21709680557251,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.013864399865269661,
+ "learning_rate": 0.001857128298614737,
+ "loss": 4.2375030517578125,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.010397658683359623,
+ "learning_rate": 0.0018558310859619048,
+ "loss": 4.231938362121582,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.00888819433748722,
+ "learning_rate": 0.0018545335987076223,
+ "loss": 4.195562839508057,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.010698067955672741,
+ "learning_rate": 0.001853235837897252,
+ "loss": 4.22847318649292,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.00987589918076992,
+ "learning_rate": 0.0018519378045763766,
+ "loss": 4.2147040367126465,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.011451782658696175,
+ "learning_rate": 0.001850639499790798,
+ "loss": 4.179263114929199,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.011224704794585705,
+ "learning_rate": 0.0018493409245865368,
+ "loss": 4.23030948638916,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.00996861606836319,
+ "learning_rate": 0.0018480420800098313,
+ "loss": 4.194283962249756,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.012143119238317013,
+ "learning_rate": 0.0018467429671071374,
+ "loss": 4.221704959869385,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.013126783072948456,
+ "learning_rate": 0.0018454435869251268,
+ "loss": 4.187748908996582,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.013031953014433384,
+ "learning_rate": 0.001844143940510687,
+ "loss": 4.248250961303711,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.012868993915617466,
+ "learning_rate": 0.001842844028910919,
+ "loss": 4.20161247253418,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.010102041065692902,
+ "learning_rate": 0.0018415438531731395,
+ "loss": 4.213654518127441,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.009599820710718632,
+ "learning_rate": 0.0018402434143448747,
+ "loss": 4.206313133239746,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.011536866426467896,
+ "learning_rate": 0.0018389427134738658,
+ "loss": 4.237072944641113,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.009558089077472687,
+ "learning_rate": 0.0018376417516080636,
+ "loss": 4.182195663452148,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01056219544261694,
+ "learning_rate": 0.0018363405297956302,
+ "loss": 4.232585906982422,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.012673584744334221,
+ "learning_rate": 0.0018350390490849354,
+ "loss": 4.228830337524414,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.011215884238481522,
+ "learning_rate": 0.0018337373105245598,
+ "loss": 4.216056823730469,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.011281915940344334,
+ "learning_rate": 0.00183243531516329,
+ "loss": 4.233404159545898,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.010381227359175682,
+ "learning_rate": 0.0018311330640501204,
+ "loss": 4.182013034820557,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.00900198519229889,
+ "learning_rate": 0.0018298305582342513,
+ "loss": 4.215488433837891,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.010198601521551609,
+ "learning_rate": 0.0018285277987650876,
+ "loss": 4.19389533996582,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.009965402074158192,
+ "learning_rate": 0.0018272247866922401,
+ "loss": 4.218005657196045,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.009299935773015022,
+ "learning_rate": 0.0018259215230655202,
+ "loss": 4.223684310913086,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.008977951481938362,
+ "learning_rate": 0.0018246180089349462,
+ "loss": 4.215778350830078,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.009261307306587696,
+ "learning_rate": 0.0018233142453507344,
+ "loss": 4.215066909790039,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.008322088047862053,
+ "learning_rate": 0.0018220102333633044,
+ "loss": 4.2188005447387695,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.008716144599020481,
+ "learning_rate": 0.0018207059740232739,
+ "loss": 4.187867641448975,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.008897020481526852,
+ "learning_rate": 0.0018194014683814625,
+ "loss": 4.218034267425537,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.007298174314200878,
+ "learning_rate": 0.0018180967174888857,
+ "loss": 4.2017011642456055,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.008491789922118187,
+ "learning_rate": 0.0018167917223967582,
+ "loss": 4.239873886108398,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.008413378149271011,
+ "learning_rate": 0.0018154864841564908,
+ "loss": 4.234771251678467,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.008141471073031425,
+ "learning_rate": 0.0018141810038196909,
+ "loss": 4.185944080352783,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.009480944834649563,
+ "learning_rate": 0.0018128752824381603,
+ "loss": 4.206586837768555,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.0106047997251153,
+ "learning_rate": 0.001811569321063894,
+ "loss": 4.235287666320801,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.010663514956831932,
+ "learning_rate": 0.0018102631207490838,
+ "loss": 4.204899311065674,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.011473523452877998,
+ "learning_rate": 0.001808956682546109,
+ "loss": 4.253668308258057,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01165992021560669,
+ "learning_rate": 0.0018076500075075456,
+ "loss": 4.209845542907715,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.011692410334944725,
+ "learning_rate": 0.0018063430966861564,
+ "loss": 4.250368595123291,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01061506848782301,
+ "learning_rate": 0.0018050359511348965,
+ "loss": 4.193821907043457,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.010510941036045551,
+ "learning_rate": 0.0018037285719069097,
+ "loss": 4.254327774047852,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.011587041430175304,
+ "learning_rate": 0.001802420960055528,
+ "loss": 4.2838945388793945,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.01257420051842928,
+ "learning_rate": 0.0018011131166342693,
+ "loss": 4.258241653442383,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.00936233438551426,
+ "learning_rate": 0.0017998050426968407,
+ "loss": 4.201854705810547,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.010093127377331257,
+ "learning_rate": 0.0017984967392971338,
+ "loss": 4.193051338195801,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.010032009333372116,
+ "learning_rate": 0.0017971882074892236,
+ "loss": 4.230212211608887,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.009455587714910507,
+ "learning_rate": 0.0017958794483273726,
+ "loss": 4.2157673835754395,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.009549238719046116,
+ "learning_rate": 0.001794570462866023,
+ "loss": 4.260272026062012,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.011059699580073357,
+ "learning_rate": 0.0017932612521598013,
+ "loss": 4.214818000793457,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.011653627268970013,
+ "learning_rate": 0.0017919518172635146,
+ "loss": 4.205112457275391,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.012438314035534859,
+ "learning_rate": 0.001790642159232151,
+ "loss": 4.179045677185059,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.011234673671424389,
+ "learning_rate": 0.0017893322791208784,
+ "loss": 4.239317417144775,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.010923845693469048,
+ "learning_rate": 0.0017880221779850434,
+ "loss": 4.230571269989014,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.009219309315085411,
+ "learning_rate": 0.001786711856880171,
+ "loss": 4.229068756103516,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.009849398396909237,
+ "learning_rate": 0.0017854013168619638,
+ "loss": 4.254772186279297,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.01117763202637434,
+ "learning_rate": 0.001784090558986299,
+ "loss": 4.252647399902344,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.010664612054824829,
+ "learning_rate": 0.0017827795843092316,
+ "loss": 4.204596519470215,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.00847950391471386,
+ "learning_rate": 0.0017814683938869895,
+ "loss": 4.211061954498291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.007305726408958435,
+ "learning_rate": 0.0017801569887759753,
+ "loss": 4.181221008300781,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.007970666512846947,
+ "learning_rate": 0.0017788453700327653,
+ "loss": 4.237855434417725,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.008873614482581615,
+ "learning_rate": 0.0017775335387141054,
+ "loss": 4.229188919067383,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.009691286832094193,
+ "learning_rate": 0.0017762214958769156,
+ "loss": 4.205491542816162,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.009360672906041145,
+ "learning_rate": 0.0017749092425782852,
+ "loss": 4.233466625213623,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.009895884431898594,
+ "learning_rate": 0.0017735967798754723,
+ "loss": 4.187763214111328,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.535672187805176,
+ "eval_runtime": 53.1736,
+ "eval_samples_per_second": 45.925,
+ "eval_steps_per_second": 1.448,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.00966129545122385,
+ "learning_rate": 0.0017722841088259045,
+ "loss": 4.1451735496521,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.010733380913734436,
+ "learning_rate": 0.0017709712304871774,
+ "loss": 4.1616363525390625,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.0117328567430377,
+ "learning_rate": 0.0017696581459170535,
+ "loss": 4.126331806182861,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.012640298344194889,
+ "learning_rate": 0.0017683448561734607,
+ "loss": 4.163730621337891,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.013074076734483242,
+ "learning_rate": 0.0017670313623144934,
+ "loss": 4.199502944946289,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.01132719125598669,
+ "learning_rate": 0.0017657176653984097,
+ "loss": 4.143380165100098,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.010008333250880241,
+ "learning_rate": 0.0017644037664836319,
+ "loss": 4.170320510864258,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.011090578511357307,
+ "learning_rate": 0.0017630896666287438,
+ "loss": 4.1571478843688965,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.011047456413507462,
+ "learning_rate": 0.0017617753668924924,
+ "loss": 4.101892948150635,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.010098653845489025,
+ "learning_rate": 0.0017604608683337857,
+ "loss": 4.154077529907227,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.010499860160052776,
+ "learning_rate": 0.00175914617201169,
+ "loss": 4.116868019104004,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.010939318686723709,
+ "learning_rate": 0.0017578312789854338,
+ "loss": 4.159144878387451,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.010622999630868435,
+ "learning_rate": 0.0017565161903144023,
+ "loss": 4.171411991119385,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.0108700692653656,
+ "learning_rate": 0.0017552009070581386,
+ "loss": 4.100748062133789,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.009794988669455051,
+ "learning_rate": 0.0017538854302763423,
+ "loss": 4.185161590576172,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.007750354241579771,
+ "learning_rate": 0.0017525697610288702,
+ "loss": 4.149083137512207,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.009923134930431843,
+ "learning_rate": 0.001751253900375732,
+ "loss": 4.205166816711426,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.010811141692101955,
+ "learning_rate": 0.0017499378493770936,
+ "loss": 4.131865978240967,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.010868688113987446,
+ "learning_rate": 0.0017486216090932733,
+ "loss": 4.148791790008545,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.011224126443266869,
+ "learning_rate": 0.001747305180584743,
+ "loss": 4.152503490447998,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.013262554071843624,
+ "learning_rate": 0.001745988564912124,
+ "loss": 4.164926528930664,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.013089599087834358,
+ "learning_rate": 0.001744671763136191,
+ "loss": 4.1626458168029785,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.01327646616846323,
+ "learning_rate": 0.0017433547763178665,
+ "loss": 4.189356803894043,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.014539042487740517,
+ "learning_rate": 0.0017420376055182233,
+ "loss": 4.176387786865234,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.013269098475575447,
+ "learning_rate": 0.001740720251798482,
+ "loss": 4.180716514587402,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.010922391898930073,
+ "learning_rate": 0.0017394027162200105,
+ "loss": 4.163928031921387,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.013355271890759468,
+ "learning_rate": 0.0017380849998443243,
+ "loss": 4.150012016296387,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.014149201102554798,
+ "learning_rate": 0.0017367671037330828,
+ "loss": 4.173713207244873,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01354879979044199,
+ "learning_rate": 0.0017354490289480918,
+ "loss": 4.150563716888428,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.013788704760372639,
+ "learning_rate": 0.0017341307765512998,
+ "loss": 4.213858604431152,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.011597222648561,
+ "learning_rate": 0.0017328123476047994,
+ "loss": 4.166989326477051,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.01055675558745861,
+ "learning_rate": 0.001731493743170824,
+ "loss": 4.172226428985596,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.011600397527217865,
+ "learning_rate": 0.0017301749643117506,
+ "loss": 4.1656951904296875,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.01114651933312416,
+ "learning_rate": 0.0017288560120900943,
+ "loss": 4.189825057983398,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.009811583906412125,
+ "learning_rate": 0.0017275368875685126,
+ "loss": 4.181164741516113,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.009758968837559223,
+ "learning_rate": 0.0017262175918097992,
+ "loss": 4.163188934326172,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.00999707542359829,
+ "learning_rate": 0.001724898125876887,
+ "loss": 4.184117317199707,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.009528180584311485,
+ "learning_rate": 0.001723578490832846,
+ "loss": 4.16407585144043,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.009717358276247978,
+ "learning_rate": 0.0017222586877408826,
+ "loss": 4.209892272949219,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.010909050703048706,
+ "learning_rate": 0.0017209387176643377,
+ "loss": 4.19977331161499,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.011234750039875507,
+ "learning_rate": 0.0017196185816666875,
+ "loss": 4.19363260269165,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.010300721041858196,
+ "learning_rate": 0.0017182982808115425,
+ "loss": 4.142427921295166,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.00941646471619606,
+ "learning_rate": 0.001716977816162644,
+ "loss": 4.159372329711914,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.009304054081439972,
+ "learning_rate": 0.0017156571887838679,
+ "loss": 4.207922458648682,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.009496180340647697,
+ "learning_rate": 0.0017143363997392184,
+ "loss": 4.137894630432129,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.009257436729967594,
+ "learning_rate": 0.0017130154500928326,
+ "loss": 4.198753833770752,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.00900586973875761,
+ "learning_rate": 0.0017116943409089746,
+ "loss": 4.160684585571289,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.009398494847118855,
+ "learning_rate": 0.00171037307325204,
+ "loss": 4.1819353103637695,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.008633013814687729,
+ "learning_rate": 0.0017090516481865483,
+ "loss": 4.176833152770996,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.00898425467312336,
+ "learning_rate": 0.0017077300667771484,
+ "loss": 4.190657138824463,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.008182398043572903,
+ "learning_rate": 0.0017064083300886155,
+ "loss": 4.173429489135742,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.009589143097400665,
+ "learning_rate": 0.0017050864391858478,
+ "loss": 4.1559014320373535,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.010304383002221584,
+ "learning_rate": 0.0017037643951338698,
+ "loss": 4.155453681945801,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.009943748824298382,
+ "learning_rate": 0.0017024421989978275,
+ "loss": 4.119734287261963,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.009238033555448055,
+ "learning_rate": 0.001701119851842992,
+ "loss": 4.188974857330322,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.008058690465986729,
+ "learning_rate": 0.0016997973547347531,
+ "loss": 4.19934606552124,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.008485953323543072,
+ "learning_rate": 0.0016984747087386233,
+ "loss": 4.165615558624268,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.008476605638861656,
+ "learning_rate": 0.0016971519149202347,
+ "loss": 4.21021842956543,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.008588708005845547,
+ "learning_rate": 0.0016958289743453385,
+ "loss": 4.208064556121826,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.009223178029060364,
+ "learning_rate": 0.0016945058880798037,
+ "loss": 4.158413887023926,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.009196167811751366,
+ "learning_rate": 0.0016931826571896166,
+ "loss": 4.203187465667725,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.009488921612501144,
+ "learning_rate": 0.001691859282740882,
+ "loss": 4.178227424621582,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.009253161028027534,
+ "learning_rate": 0.0016905357657998167,
+ "loss": 4.131230354309082,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.010069048963487148,
+ "learning_rate": 0.0016892121074327559,
+ "loss": 4.210734844207764,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.009945095516741276,
+ "learning_rate": 0.001687888308706147,
+ "loss": 4.184042930603027,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.010483640246093273,
+ "learning_rate": 0.0016865643706865505,
+ "loss": 4.156919479370117,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.011560715734958649,
+ "learning_rate": 0.0016852402944406391,
+ "loss": 4.185520172119141,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.010001277551054955,
+ "learning_rate": 0.0016839160810351978,
+ "loss": 4.226828575134277,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.010335934348404408,
+ "learning_rate": 0.0016825917315371202,
+ "loss": 4.183531284332275,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.011322776786983013,
+ "learning_rate": 0.001681267247013412,
+ "loss": 4.179635047912598,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.009663194417953491,
+ "learning_rate": 0.001679942628531186,
+ "loss": 4.165654182434082,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.010590246878564358,
+ "learning_rate": 0.0016786178771576643,
+ "loss": 4.1539459228515625,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.511386394500732,
+ "eval_runtime": 51.7751,
+ "eval_samples_per_second": 47.166,
+ "eval_steps_per_second": 1.487,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.01254354789853096,
+ "learning_rate": 0.0016772929939601728,
+ "loss": 4.104002475738525,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.01300987508147955,
+ "learning_rate": 0.0016759679800061483,
+ "loss": 4.127395153045654,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.013325816951692104,
+ "learning_rate": 0.0016746428363631299,
+ "loss": 4.131229400634766,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.014302870258688927,
+ "learning_rate": 0.001673317564098761,
+ "loss": 4.079358100891113,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.013776068575680256,
+ "learning_rate": 0.0016719921642807905,
+ "loss": 4.132065773010254,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.010788560844957829,
+ "learning_rate": 0.0016706666379770689,
+ "loss": 4.134626388549805,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.009766630828380585,
+ "learning_rate": 0.0016693409862555493,
+ "loss": 4.1645426750183105,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.011760029010474682,
+ "learning_rate": 0.0016680152101842842,
+ "loss": 4.110664367675781,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.011047836393117905,
+ "learning_rate": 0.0016666893108314288,
+ "loss": 4.129222869873047,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.009989139623939991,
+ "learning_rate": 0.0016653632892652344,
+ "loss": 4.076835632324219,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.010033783502876759,
+ "learning_rate": 0.001664037146554054,
+ "loss": 4.11555290222168,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.00996427796781063,
+ "learning_rate": 0.001662710883766337,
+ "loss": 4.131463527679443,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.01044023409485817,
+ "learning_rate": 0.0016613845019706288,
+ "loss": 4.098750114440918,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.010033872909843922,
+ "learning_rate": 0.0016600580022355712,
+ "loss": 4.161447048187256,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.009329567663371563,
+ "learning_rate": 0.0016587313856299012,
+ "loss": 4.146491527557373,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.010061589069664478,
+ "learning_rate": 0.0016574046532224504,
+ "loss": 4.107517719268799,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.009491910226643085,
+ "learning_rate": 0.0016560778060821424,
+ "loss": 4.098226547241211,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.008529345504939556,
+ "learning_rate": 0.0016547508452779943,
+ "loss": 4.1247711181640625,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.00936342403292656,
+ "learning_rate": 0.0016534237718791136,
+ "loss": 4.120573043823242,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.010255641303956509,
+ "learning_rate": 0.0016520965869547008,
+ "loss": 4.089067459106445,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.011638758704066277,
+ "learning_rate": 0.001650769291574044,
+ "loss": 4.074847221374512,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.014707857742905617,
+ "learning_rate": 0.0016494418868065214,
+ "loss": 4.153377532958984,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.01650111749768257,
+ "learning_rate": 0.0016481143737215986,
+ "loss": 4.134062767028809,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.014833372086286545,
+ "learning_rate": 0.00164678675338883,
+ "loss": 4.1344828605651855,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.012257886119186878,
+ "learning_rate": 0.0016454590268778538,
+ "loss": 4.098813533782959,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.013508322648704052,
+ "learning_rate": 0.0016441311952583959,
+ "loss": 4.106297492980957,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.01256900280714035,
+ "learning_rate": 0.001642803259600267,
+ "loss": 4.143631935119629,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.009885406121611595,
+ "learning_rate": 0.00164147522097336,
+ "loss": 4.120885848999023,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.012004674412310123,
+ "learning_rate": 0.0016401470804476518,
+ "loss": 4.089243412017822,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01224028691649437,
+ "learning_rate": 0.0016388188390932012,
+ "loss": 4.077765464782715,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.012511290609836578,
+ "learning_rate": 0.001637490497980149,
+ "loss": 4.161346435546875,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.010701716877520084,
+ "learning_rate": 0.0016361620581787143,
+ "loss": 4.10698938369751,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.010713113471865654,
+ "learning_rate": 0.001634833520759198,
+ "loss": 4.114283561706543,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.011018907651305199,
+ "learning_rate": 0.001633504886791978,
+ "loss": 4.134284019470215,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.011279584839940071,
+ "learning_rate": 0.0016321761573475117,
+ "loss": 4.1456804275512695,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.010642931796610355,
+ "learning_rate": 0.001630847333496332,
+ "loss": 4.160918235778809,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.011069678701460361,
+ "learning_rate": 0.0016295184163090482,
+ "loss": 4.136838912963867,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.011007166467607021,
+ "learning_rate": 0.0016281894068563443,
+ "loss": 4.129153728485107,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.010055703110992908,
+ "learning_rate": 0.0016268603062089802,
+ "loss": 4.138665676116943,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.009839230217039585,
+ "learning_rate": 0.0016255311154377875,
+ "loss": 4.1366496086120605,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.011191432364284992,
+ "learning_rate": 0.0016242018356136714,
+ "loss": 4.129880905151367,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.012130341492593288,
+ "learning_rate": 0.0016228724678076095,
+ "loss": 4.1425981521606445,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.012707795947790146,
+ "learning_rate": 0.0016215430130906482,
+ "loss": 4.115452766418457,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.012707809917628765,
+ "learning_rate": 0.001620213472533906,
+ "loss": 4.122758865356445,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.01182397548109293,
+ "learning_rate": 0.001618883847208569,
+ "loss": 4.140275955200195,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.010865340940654278,
+ "learning_rate": 0.001617554138185893,
+ "loss": 4.137759208679199,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.010973592288792133,
+ "learning_rate": 0.0016162243465371998,
+ "loss": 4.138947010040283,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.011247226968407631,
+ "learning_rate": 0.001614894473333879,
+ "loss": 4.115516662597656,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.010837487876415253,
+ "learning_rate": 0.0016135645196473853,
+ "loss": 4.160427093505859,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.009096396155655384,
+ "learning_rate": 0.0016122344865492385,
+ "loss": 4.1200032234191895,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.010671247728168964,
+ "learning_rate": 0.0016109043751110223,
+ "loss": 4.097919464111328,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.0107562355697155,
+ "learning_rate": 0.0016095741864043831,
+ "loss": 4.096290588378906,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.008909327909350395,
+ "learning_rate": 0.0016082439215010305,
+ "loss": 4.171017646789551,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.010305124334990978,
+ "learning_rate": 0.0016069135814727343,
+ "loss": 4.1450276374816895,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.011235613375902176,
+ "learning_rate": 0.0016055831673913255,
+ "loss": 4.134704113006592,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.00825632456690073,
+ "learning_rate": 0.0016042526803286949,
+ "loss": 4.107443332672119,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.009056378155946732,
+ "learning_rate": 0.0016029221213567921,
+ "loss": 4.132656097412109,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.008835164830088615,
+ "learning_rate": 0.001601591491547624,
+ "loss": 4.124729633331299,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.00915333442389965,
+ "learning_rate": 0.0016002607919732557,
+ "loss": 4.119500160217285,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.008532539941370487,
+ "learning_rate": 0.0015989300237058066,
+ "loss": 4.141661643981934,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.009952311404049397,
+ "learning_rate": 0.0015975991878174533,
+ "loss": 4.140929222106934,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.008920297026634216,
+ "learning_rate": 0.0015962682853804268,
+ "loss": 4.155877113342285,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.008770049549639225,
+ "learning_rate": 0.0015949373174670104,
+ "loss": 4.126394271850586,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.0078084105625748634,
+ "learning_rate": 0.0015936062851495411,
+ "loss": 4.128876209259033,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.00816641841083765,
+ "learning_rate": 0.0015922751895004078,
+ "loss": 4.137410640716553,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.008895459584891796,
+ "learning_rate": 0.0015909440315920505,
+ "loss": 4.07957649230957,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.008551528677344322,
+ "learning_rate": 0.001589612812496958,
+ "loss": 4.104213714599609,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.008325488306581974,
+ "learning_rate": 0.0015882815332876707,
+ "loss": 4.1584792137146,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.009341313503682613,
+ "learning_rate": 0.0015869501950367744,
+ "loss": 4.1417436599731445,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.00999050959944725,
+ "learning_rate": 0.0015856187988169067,
+ "loss": 4.136049270629883,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.008889932185411453,
+ "learning_rate": 0.0015842873457007484,
+ "loss": 4.16257381439209,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.009109573438763618,
+ "learning_rate": 0.0015829558367610268,
+ "loss": 4.171441555023193,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.485942840576172,
+ "eval_runtime": 52.1051,
+ "eval_samples_per_second": 46.867,
+ "eval_steps_per_second": 1.478,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.009881424717605114,
+ "learning_rate": 0.001581624273070515,
+ "loss": 4.059608459472656,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.011156721040606499,
+ "learning_rate": 0.0015802926557020294,
+ "loss": 4.047014236450195,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.012326992116868496,
+ "learning_rate": 0.0015789609857284309,
+ "loss": 4.06541109085083,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.01332208700478077,
+ "learning_rate": 0.0015776292642226213,
+ "loss": 4.042714595794678,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.012720128521323204,
+ "learning_rate": 0.0015762974922575446,
+ "loss": 4.098389625549316,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.010723215527832508,
+ "learning_rate": 0.0015749656709061854,
+ "loss": 4.059185981750488,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.010605281218886375,
+ "learning_rate": 0.001573633801241569,
+ "loss": 4.099538326263428,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.01101530622690916,
+ "learning_rate": 0.0015723018843367576,
+ "loss": 4.095674514770508,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.01135838869959116,
+ "learning_rate": 0.0015709699212648532,
+ "loss": 4.070535659790039,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.012119271792471409,
+ "learning_rate": 0.0015696379130989936,
+ "loss": 4.060605049133301,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.012783961370587349,
+ "learning_rate": 0.001568305860912355,
+ "loss": 4.075931072235107,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.011486220173537731,
+ "learning_rate": 0.0015669737657781461,
+ "loss": 4.075855255126953,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.01037552859634161,
+ "learning_rate": 0.0015656416287696136,
+ "loss": 4.087729454040527,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.009802743792533875,
+ "learning_rate": 0.0015643094509600354,
+ "loss": 4.0855607986450195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.010804438032209873,
+ "learning_rate": 0.001562977233422723,
+ "loss": 4.064528465270996,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.009505728259682655,
+ "learning_rate": 0.0015616449772310206,
+ "loss": 4.036775588989258,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.010634485632181168,
+ "learning_rate": 0.0015603126834583024,
+ "loss": 4.069923400878906,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.011392096057534218,
+ "learning_rate": 0.0015589803531779736,
+ "loss": 4.0812177658081055,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.011851722374558449,
+ "learning_rate": 0.0015576479874634683,
+ "loss": 4.09183406829834,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.011205444112420082,
+ "learning_rate": 0.0015563155873882502,
+ "loss": 4.07825231552124,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.009676349349319935,
+ "learning_rate": 0.0015549831540258093,
+ "loss": 4.052826404571533,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.009376165457069874,
+ "learning_rate": 0.0015536506884496638,
+ "loss": 4.02100133895874,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.00936955213546753,
+ "learning_rate": 0.0015523181917333564,
+ "loss": 4.049526214599609,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.00937784556299448,
+ "learning_rate": 0.0015509856649504564,
+ "loss": 4.096356391906738,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.010014008730649948,
+ "learning_rate": 0.0015496531091745555,
+ "loss": 4.03254508972168,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.009909634478390217,
+ "learning_rate": 0.0015483205254792705,
+ "loss": 4.067630767822266,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.009319501928985119,
+ "learning_rate": 0.0015469879149382399,
+ "loss": 4.094167709350586,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.008634721860289574,
+ "learning_rate": 0.0015456552786251244,
+ "loss": 4.122404098510742,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.009412454441189766,
+ "learning_rate": 0.0015443226176136049,
+ "loss": 4.081546783447266,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.010102108120918274,
+ "learning_rate": 0.0015429899329773814,
+ "loss": 4.049839019775391,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.010921654291450977,
+ "learning_rate": 0.0015416572257901746,
+ "loss": 4.097097396850586,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.010607713833451271,
+ "learning_rate": 0.0015403244971257226,
+ "loss": 4.084481239318848,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.011887960135936737,
+ "learning_rate": 0.0015389917480577806,
+ "loss": 4.074968338012695,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01089888159185648,
+ "learning_rate": 0.0015376589796601203,
+ "loss": 4.062132835388184,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.009934183210134506,
+ "learning_rate": 0.0015363261930065298,
+ "loss": 4.128092288970947,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.00878534372895956,
+ "learning_rate": 0.0015349933891708105,
+ "loss": 4.069051742553711,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.009928155690431595,
+ "learning_rate": 0.0015336605692267788,
+ "loss": 4.073912620544434,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.011091411113739014,
+ "learning_rate": 0.0015323277342482637,
+ "loss": 4.125944137573242,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.010314172133803368,
+ "learning_rate": 0.0015309948853091055,
+ "loss": 4.062620162963867,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.010423967614769936,
+ "learning_rate": 0.001529662023483158,
+ "loss": 4.101737976074219,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.009822997264564037,
+ "learning_rate": 0.0015283291498442817,
+ "loss": 4.064520359039307,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.009757095016539097,
+ "learning_rate": 0.001526996265466351,
+ "loss": 4.100624084472656,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.008650688454508781,
+ "learning_rate": 0.001525663371423246,
+ "loss": 4.109685897827148,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.007378234062343836,
+ "learning_rate": 0.0015243304687888553,
+ "loss": 4.044726848602295,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.008194336667656898,
+ "learning_rate": 0.001522997558637074,
+ "loss": 4.097923278808594,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.009764892049133778,
+ "learning_rate": 0.0015216646420418048,
+ "loss": 4.102148056030273,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.009242868050932884,
+ "learning_rate": 0.001520331720076953,
+ "loss": 4.068563461303711,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.010409324429929256,
+ "learning_rate": 0.001518998793816431,
+ "loss": 4.065944194793701,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.009215258061885834,
+ "learning_rate": 0.0015176658643341527,
+ "loss": 4.10223388671875,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.008065393194556236,
+ "learning_rate": 0.0015163329327040366,
+ "loss": 4.099432945251465,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.008498288691043854,
+ "learning_rate": 0.001515,
+ "loss": 4.109854698181152,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.008919150568544865,
+ "learning_rate": 0.0015136670672959633,
+ "loss": 4.128247261047363,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.009104429744184017,
+ "learning_rate": 0.0015123341356658474,
+ "loss": 4.068622589111328,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.008558365516364574,
+ "learning_rate": 0.0015110012061835692,
+ "loss": 4.114214897155762,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.007865076884627342,
+ "learning_rate": 0.0015096682799230477,
+ "loss": 4.11155366897583,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.007848534733057022,
+ "learning_rate": 0.0015083353579581958,
+ "loss": 4.045694351196289,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.008238030597567558,
+ "learning_rate": 0.0015070024413629264,
+ "loss": 4.093930721282959,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.007536398712545633,
+ "learning_rate": 0.0015056695312111448,
+ "loss": 4.106235504150391,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.008156892843544483,
+ "learning_rate": 0.001504336628576754,
+ "loss": 4.13060998916626,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.008144439198076725,
+ "learning_rate": 0.0015030037345336487,
+ "loss": 4.091304779052734,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.007757358253002167,
+ "learning_rate": 0.001501670850155718,
+ "loss": 4.069276809692383,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.00860623549669981,
+ "learning_rate": 0.0015003379765168425,
+ "loss": 4.107658386230469,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.008614457212388515,
+ "learning_rate": 0.0014990051146908949,
+ "loss": 4.106705665588379,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.009120592847466469,
+ "learning_rate": 0.001497672265751737,
+ "loss": 4.0891923904418945,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.009446980431675911,
+ "learning_rate": 0.0014963394307732213,
+ "loss": 4.098847389221191,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.010494766756892204,
+ "learning_rate": 0.0014950066108291894,
+ "loss": 4.115325450897217,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.010057658888399601,
+ "learning_rate": 0.0014936738069934706,
+ "loss": 4.099984169006348,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.009071771055459976,
+ "learning_rate": 0.0014923410203398796,
+ "loss": 4.091543674468994,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.009937075898051262,
+ "learning_rate": 0.00149100825194222,
+ "loss": 4.086492538452148,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.010222986340522766,
+ "learning_rate": 0.0014896755028742778,
+ "loss": 4.063841819763184,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.010513762012124062,
+ "learning_rate": 0.001488342774209826,
+ "loss": 4.11991548538208,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.009530222974717617,
+ "learning_rate": 0.0014870100670226192,
+ "loss": 4.043334007263184,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.467459678649902,
+ "eval_runtime": 54.1628,
+ "eval_samples_per_second": 45.086,
+ "eval_steps_per_second": 1.422,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.010274187661707401,
+ "learning_rate": 0.0014856773823863955,
+ "loss": 4.049561500549316,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.011448442004621029,
+ "learning_rate": 0.001484344721374876,
+ "loss": 4.019484996795654,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.011361192911863327,
+ "learning_rate": 0.0014830120850617599,
+ "loss": 4.0429887771606445,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.010250156745314598,
+ "learning_rate": 0.0014816794745207299,
+ "loss": 4.012682914733887,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.010024400427937508,
+ "learning_rate": 0.0014803468908254449,
+ "loss": 3.979409694671631,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.01153535582125187,
+ "learning_rate": 0.0014790143350495446,
+ "loss": 4.045780181884766,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.012632817961275578,
+ "learning_rate": 0.001477681808266644,
+ "loss": 4.057641983032227,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.013758293353021145,
+ "learning_rate": 0.0014763493115503366,
+ "loss": 4.061607360839844,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.014041583053767681,
+ "learning_rate": 0.001475016845974191,
+ "loss": 4.035029411315918,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.013629456982016563,
+ "learning_rate": 0.0014736844126117497,
+ "loss": 4.062169075012207,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.011023213155567646,
+ "learning_rate": 0.0014723520125365323,
+ "loss": 4.02731990814209,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.01001750584691763,
+ "learning_rate": 0.0014710196468220266,
+ "loss": 4.019882678985596,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.008321646600961685,
+ "learning_rate": 0.001469687316541698,
+ "loss": 4.005446434020996,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.008935946971178055,
+ "learning_rate": 0.0014683550227689796,
+ "loss": 4.017102241516113,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.009137130342423916,
+ "learning_rate": 0.001467022766577277,
+ "loss": 4.024102210998535,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.00805022194981575,
+ "learning_rate": 0.001465690549039965,
+ "loss": 3.9935250282287598,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.009204058907926083,
+ "learning_rate": 0.0014643583712303865,
+ "loss": 4.042818546295166,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.009812815114855766,
+ "learning_rate": 0.001463026234221854,
+ "loss": 4.0252485275268555,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.009955782443284988,
+ "learning_rate": 0.0014616941390876454,
+ "loss": 3.9839768409729004,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.009374645538628101,
+ "learning_rate": 0.001460362086901007,
+ "loss": 4.040245056152344,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.010008739307522774,
+ "learning_rate": 0.0014590300787351471,
+ "loss": 4.0306596755981445,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.009014200419187546,
+ "learning_rate": 0.001457698115663243,
+ "loss": 4.028713226318359,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.008381523191928864,
+ "learning_rate": 0.0014563661987584313,
+ "loss": 4.024718284606934,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.008627948351204395,
+ "learning_rate": 0.0014550343290938143,
+ "loss": 4.02303409576416,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.010257705114781857,
+ "learning_rate": 0.0014537025077424556,
+ "loss": 4.061058044433594,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.010759209282696247,
+ "learning_rate": 0.0014523707357773788,
+ "loss": 4.064718246459961,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.009966139681637287,
+ "learning_rate": 0.0014510390142715697,
+ "loss": 3.9919447898864746,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.00997584406286478,
+ "learning_rate": 0.001449707344297971,
+ "loss": 4.0431599617004395,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.010846944525837898,
+ "learning_rate": 0.0014483757269294858,
+ "loss": 4.0222697257995605,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.00991352554410696,
+ "learning_rate": 0.0014470441632389735,
+ "loss": 4.059438705444336,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.008994397707283497,
+ "learning_rate": 0.0014457126542992518,
+ "loss": 4.021494388580322,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.009567267261445522,
+ "learning_rate": 0.0014443812011830934,
+ "loss": 4.070810317993164,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.009199011139571667,
+ "learning_rate": 0.0014430498049632255,
+ "loss": 4.020938873291016,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.010440350510179996,
+ "learning_rate": 0.0014417184667123303,
+ "loss": 4.062529563903809,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.009962482377886772,
+ "learning_rate": 0.0014403871875030424,
+ "loss": 4.0687150955200195,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.009389379993081093,
+ "learning_rate": 0.0014390559684079502,
+ "loss": 4.055185794830322,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.011469941586256027,
+ "learning_rate": 0.0014377248104995924,
+ "loss": 4.041023254394531,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.011034650728106499,
+ "learning_rate": 0.0014363937148504588,
+ "loss": 4.054254531860352,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.010651151649653912,
+ "learning_rate": 0.0014350626825329897,
+ "loss": 4.037753582000732,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.010747930034995079,
+ "learning_rate": 0.0014337317146195733,
+ "loss": 4.059938907623291,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.00987576600164175,
+ "learning_rate": 0.0014324008121825468,
+ "loss": 4.075366973876953,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.008504371158778667,
+ "learning_rate": 0.0014310699762941936,
+ "loss": 4.026755332946777,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.009445052593946457,
+ "learning_rate": 0.001429739208026745,
+ "loss": 4.034431457519531,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.010560023598372936,
+ "learning_rate": 0.0014284085084523763,
+ "loss": 4.101158618927002,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.009920580312609673,
+ "learning_rate": 0.0014270778786432078,
+ "loss": 4.041321754455566,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.009854268282651901,
+ "learning_rate": 0.0014257473196713053,
+ "loss": 4.049882411956787,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.009825135581195354,
+ "learning_rate": 0.0014244168326086745,
+ "loss": 4.067440986633301,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.010506784543395042,
+ "learning_rate": 0.0014230864185272663,
+ "loss": 4.042476654052734,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.009527767077088356,
+ "learning_rate": 0.0014217560784989696,
+ "loss": 4.055248260498047,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.01013826485723257,
+ "learning_rate": 0.001420425813595617,
+ "loss": 4.05246639251709,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.011020111851394176,
+ "learning_rate": 0.0014190956248889778,
+ "loss": 4.032987117767334,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.010647117160260677,
+ "learning_rate": 0.0014177655134507614,
+ "loss": 4.093589782714844,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.009484780952334404,
+ "learning_rate": 0.001416435480352615,
+ "loss": 4.017437934875488,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.0114199910312891,
+ "learning_rate": 0.0014151055266661211,
+ "loss": 4.080643653869629,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.011137664318084717,
+ "learning_rate": 0.0014137756534628007,
+ "loss": 4.069380760192871,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.008972682990133762,
+ "learning_rate": 0.0014124458618141073,
+ "loss": 4.051383018493652,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.010311643593013287,
+ "learning_rate": 0.0014111161527914317,
+ "loss": 4.048953533172607,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.01113453321158886,
+ "learning_rate": 0.0014097865274660944,
+ "loss": 4.063544273376465,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.01091703586280346,
+ "learning_rate": 0.0014084569869093518,
+ "loss": 4.006721496582031,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.010672449134290218,
+ "learning_rate": 0.0014071275321923906,
+ "loss": 4.020220756530762,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.009682527743279934,
+ "learning_rate": 0.001405798164386328,
+ "loss": 4.057307243347168,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.009808676317334175,
+ "learning_rate": 0.0014044688845622126,
+ "loss": 4.105412483215332,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.010351189412176609,
+ "learning_rate": 0.0014031396937910202,
+ "loss": 4.066454887390137,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.010038275271654129,
+ "learning_rate": 0.001401810593143656,
+ "loss": 4.056752681732178,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01016094908118248,
+ "learning_rate": 0.001400481583690952,
+ "loss": 4.097933292388916,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.008683547377586365,
+ "learning_rate": 0.0013991526665036679,
+ "loss": 4.0386152267456055,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.009556332603096962,
+ "learning_rate": 0.0013978238426524882,
+ "loss": 4.05835485458374,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.009597044438123703,
+ "learning_rate": 0.001396495113208022,
+ "loss": 4.052363872528076,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.010021139867603779,
+ "learning_rate": 0.0013951664792408026,
+ "loss": 4.083431243896484,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.010611358098685741,
+ "learning_rate": 0.0013938379418212863,
+ "loss": 4.042330265045166,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.009132112376391888,
+ "learning_rate": 0.001392509502019852,
+ "loss": 4.058825969696045,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.009808709844946861,
+ "learning_rate": 0.0013911811609067992,
+ "loss": 4.093010902404785,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.4473161697387695,
+ "eval_runtime": 51.7738,
+ "eval_samples_per_second": 47.167,
+ "eval_steps_per_second": 1.487,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.011859052814543247,
+ "learning_rate": 0.0013898529195523488,
+ "loss": 3.9572651386260986,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.013210238888859749,
+ "learning_rate": 0.0013885247790266406,
+ "loss": 3.9812350273132324,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.012981324456632137,
+ "learning_rate": 0.001387196740399733,
+ "loss": 3.987725019454956,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.01424382347613573,
+ "learning_rate": 0.001385868804741604,
+ "loss": 3.9719889163970947,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.0154642965644598,
+ "learning_rate": 0.0013845409731221465,
+ "loss": 3.984248161315918,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.014779271557927132,
+ "learning_rate": 0.0013832132466111706,
+ "loss": 4.002912521362305,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.012921188026666641,
+ "learning_rate": 0.0013818856262784013,
+ "loss": 3.9588499069213867,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01165325753390789,
+ "learning_rate": 0.0013805581131934792,
+ "loss": 3.993288993835449,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.008908954448997974,
+ "learning_rate": 0.0013792307084259563,
+ "loss": 3.987151861190796,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.012040507979691029,
+ "learning_rate": 0.0013779034130452991,
+ "loss": 3.9675703048706055,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.013464349322021008,
+ "learning_rate": 0.0013765762281208868,
+ "loss": 3.9901437759399414,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.009626250714063644,
+ "learning_rate": 0.0013752491547220063,
+ "loss": 3.9373281002044678,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.00897257961332798,
+ "learning_rate": 0.0013739221939178582,
+ "loss": 4.00804328918457,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.010689414106309414,
+ "learning_rate": 0.00137259534677755,
+ "loss": 4.020804405212402,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.010462156496942043,
+ "learning_rate": 0.0013712686143700988,
+ "loss": 4.028973579406738,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.009135857224464417,
+ "learning_rate": 0.001369941997764429,
+ "loss": 3.9798216819763184,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.009298899210989475,
+ "learning_rate": 0.0013686154980293713,
+ "loss": 4.018845081329346,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.00979928020387888,
+ "learning_rate": 0.0013672891162336636,
+ "loss": 3.9946775436401367,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.008175542578101158,
+ "learning_rate": 0.0013659628534459457,
+ "loss": 3.981858491897583,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.008302642032504082,
+ "learning_rate": 0.0013646367107347662,
+ "loss": 3.991128444671631,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.008516009896993637,
+ "learning_rate": 0.0013633106891685718,
+ "loss": 4.004990100860596,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.007936482317745686,
+ "learning_rate": 0.0013619847898157164,
+ "loss": 3.995734214782715,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.008317600004374981,
+ "learning_rate": 0.001360659013744451,
+ "loss": 4.040653228759766,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.008481282740831375,
+ "learning_rate": 0.001359333362022931,
+ "loss": 3.969860553741455,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.008944948203861713,
+ "learning_rate": 0.0013580078357192096,
+ "loss": 3.9908909797668457,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.008897850289940834,
+ "learning_rate": 0.001356682435901239,
+ "loss": 4.04786491394043,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.009527456946671009,
+ "learning_rate": 0.001355357163636871,
+ "loss": 4.017423629760742,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.011714582331478596,
+ "learning_rate": 0.001354032019993852,
+ "loss": 3.9990692138671875,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.011083807796239853,
+ "learning_rate": 0.0013527070060398274,
+ "loss": 3.995265483856201,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.009878288954496384,
+ "learning_rate": 0.0013513821228423363,
+ "loss": 4.0364532470703125,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.009727329947054386,
+ "learning_rate": 0.0013500573714688136,
+ "loss": 4.035953521728516,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.009482033550739288,
+ "learning_rate": 0.0013487327529865879,
+ "loss": 4.015855312347412,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.009037062525749207,
+ "learning_rate": 0.0013474082684628798,
+ "loss": 4.024164199829102,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.009094092063605785,
+ "learning_rate": 0.0013460839189648028,
+ "loss": 4.037765979766846,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.008230654522776604,
+ "learning_rate": 0.0013447597055593612,
+ "loss": 4.006317615509033,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.008491289801895618,
+ "learning_rate": 0.0013434356293134501,
+ "loss": 4.042603969573975,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.008907693438231945,
+ "learning_rate": 0.0013421116912938533,
+ "loss": 4.008098602294922,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.008897818624973297,
+ "learning_rate": 0.0013407878925672438,
+ "loss": 3.9892945289611816,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.00957555789500475,
+ "learning_rate": 0.0013394642342001832,
+ "loss": 4.046009063720703,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.009948568418622017,
+ "learning_rate": 0.0013381407172591185,
+ "loss": 3.9760642051696777,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.01014955248683691,
+ "learning_rate": 0.0013368173428103835,
+ "loss": 4.023215293884277,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.008578895591199398,
+ "learning_rate": 0.0013354941119201966,
+ "loss": 4.002819061279297,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.008729875087738037,
+ "learning_rate": 0.001334171025654662,
+ "loss": 4.0340657234191895,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.009556280449032784,
+ "learning_rate": 0.0013328480850797655,
+ "loss": 3.983076333999634,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.009733820334076881,
+ "learning_rate": 0.0013315252912613766,
+ "loss": 4.066578388214111,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.009025252424180508,
+ "learning_rate": 0.0013302026452652474,
+ "loss": 3.9980697631835938,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.008485320024192333,
+ "learning_rate": 0.001328880148157008,
+ "loss": 4.018826961517334,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.009539696387946606,
+ "learning_rate": 0.0013275578010021726,
+ "loss": 4.011737823486328,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.010881016030907631,
+ "learning_rate": 0.0013262356048661305,
+ "loss": 4.007291793823242,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.010362236760556698,
+ "learning_rate": 0.0013249135608141523,
+ "loss": 3.9929394721984863,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.010296608321368694,
+ "learning_rate": 0.0013235916699113849,
+ "loss": 4.0015130043029785,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.011211174540221691,
+ "learning_rate": 0.0013222699332228517,
+ "loss": 4.005324363708496,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.009743994101881981,
+ "learning_rate": 0.0013209483518134525,
+ "loss": 4.0278143882751465,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.008563914336264133,
+ "learning_rate": 0.0013196269267479607,
+ "loss": 3.9997801780700684,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.009078902192413807,
+ "learning_rate": 0.0013183056590910255,
+ "loss": 4.026200294494629,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.009618895128369331,
+ "learning_rate": 0.0013169845499071675,
+ "loss": 4.027050018310547,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.008920644409954548,
+ "learning_rate": 0.0013156636002607822,
+ "loss": 4.000490665435791,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.008849202655255795,
+ "learning_rate": 0.0013143428112161323,
+ "loss": 4.044985294342041,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.009303621016442776,
+ "learning_rate": 0.001313022183837356,
+ "loss": 4.046281814575195,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.01076227892190218,
+ "learning_rate": 0.0013117017191884579,
+ "loss": 4.030576705932617,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.010400841943919659,
+ "learning_rate": 0.0013103814183333127,
+ "loss": 3.980081558227539,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.009459166787564754,
+ "learning_rate": 0.001309061282335663,
+ "loss": 4.014824867248535,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.009054969064891338,
+ "learning_rate": 0.001307741312259118,
+ "loss": 4.0485944747924805,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.009003991261124611,
+ "learning_rate": 0.0013064215091671545,
+ "loss": 3.9990320205688477,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.008959448896348476,
+ "learning_rate": 0.001305101874123113,
+ "loss": 4.038620948791504,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.009102243930101395,
+ "learning_rate": 0.0013037824081902015,
+ "loss": 4.030906677246094,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.008599337190389633,
+ "learning_rate": 0.0013024631124314877,
+ "loss": 4.02648401260376,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.008771977387368679,
+ "learning_rate": 0.0013011439879099056,
+ "loss": 4.027151107788086,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.009082595817744732,
+ "learning_rate": 0.00129982503568825,
+ "loss": 4.0194478034973145,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.009025354869663715,
+ "learning_rate": 0.0012985062568291763,
+ "loss": 4.044382095336914,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.00875102635473013,
+ "learning_rate": 0.0012971876523952014,
+ "loss": 4.0257568359375,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.009312704205513,
+ "learning_rate": 0.0012958692234487006,
+ "loss": 3.9964494705200195,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.431707859039307,
+ "eval_runtime": 50.7311,
+ "eval_samples_per_second": 48.136,
+ "eval_steps_per_second": 1.518,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.00992550328373909,
+ "learning_rate": 0.0012945509710519088,
+ "loss": 3.9253792762756348,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.009920453652739525,
+ "learning_rate": 0.0012932328962669173,
+ "loss": 3.9438819885253906,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.009437218308448792,
+ "learning_rate": 0.0012919150001556756,
+ "loss": 3.960963249206543,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.009616893716156483,
+ "learning_rate": 0.0012905972837799896,
+ "loss": 3.969301223754883,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.009107022546231747,
+ "learning_rate": 0.0012892797482015185,
+ "loss": 3.9626078605651855,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.011317167431116104,
+ "learning_rate": 0.0012879623944817772,
+ "loss": 3.946483850479126,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.010536770336329937,
+ "learning_rate": 0.001286645223682134,
+ "loss": 3.959566116333008,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.009349033236503601,
+ "learning_rate": 0.0012853282368638097,
+ "loss": 3.964435338973999,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.009950080886483192,
+ "learning_rate": 0.0012840114350878762,
+ "loss": 3.9729413986206055,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.010833023115992546,
+ "learning_rate": 0.001282694819415257,
+ "loss": 4.006421089172363,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.011538525111973286,
+ "learning_rate": 0.0012813783909067266,
+ "loss": 3.9558355808258057,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.010671315714716911,
+ "learning_rate": 0.0012800621506229065,
+ "loss": 4.0052337646484375,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.010835555382072926,
+ "learning_rate": 0.0012787460996242686,
+ "loss": 3.958641767501831,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.01082367543131113,
+ "learning_rate": 0.0012774302389711304,
+ "loss": 3.9818902015686035,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.009194346144795418,
+ "learning_rate": 0.0012761145697236583,
+ "loss": 3.9688854217529297,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.007859935984015465,
+ "learning_rate": 0.001274799092941862,
+ "loss": 3.9730639457702637,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.008534743450582027,
+ "learning_rate": 0.0012734838096855978,
+ "loss": 3.975003719329834,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.009369445964694023,
+ "learning_rate": 0.0012721687210145666,
+ "loss": 3.972907304763794,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.0090147340670228,
+ "learning_rate": 0.0012708538279883098,
+ "loss": 3.964256525039673,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.007745139300823212,
+ "learning_rate": 0.0012695391316662151,
+ "loss": 3.9504833221435547,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.008920101448893547,
+ "learning_rate": 0.0012682246331075075,
+ "loss": 3.9735097885131836,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.008987226523458958,
+ "learning_rate": 0.0012669103333712568,
+ "loss": 3.9482927322387695,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.00841289572417736,
+ "learning_rate": 0.0012655962335163685,
+ "loss": 3.965221881866455,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.008016116917133331,
+ "learning_rate": 0.0012642823346015902,
+ "loss": 3.9390206336975098,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.008470152504742146,
+ "learning_rate": 0.0012629686376855068,
+ "loss": 3.9257445335388184,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.008637234568595886,
+ "learning_rate": 0.0012616551438265394,
+ "loss": 3.9999356269836426,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.008036923594772816,
+ "learning_rate": 0.0012603418540829473,
+ "loss": 3.9755172729492188,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.008908843621611595,
+ "learning_rate": 0.0012590287695128232,
+ "loss": 3.9248545169830322,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.008743193000555038,
+ "learning_rate": 0.001257715891174096,
+ "loss": 3.9619390964508057,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.00834199320524931,
+ "learning_rate": 0.001256403220124528,
+ "loss": 3.961569309234619,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.009148893877863884,
+ "learning_rate": 0.001255090757421715,
+ "loss": 3.972104549407959,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.00925084762275219,
+ "learning_rate": 0.0012537785041230843,
+ "loss": 4.000775337219238,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.008659154176712036,
+ "learning_rate": 0.0012524664612858947,
+ "loss": 3.954564332962036,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.007787859998643398,
+ "learning_rate": 0.0012511546299672355,
+ "loss": 3.9498860836029053,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.00817918125540018,
+ "learning_rate": 0.0012498430112240246,
+ "loss": 4.003897666931152,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.009282548911869526,
+ "learning_rate": 0.001248531606113011,
+ "loss": 3.9621925354003906,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.009375610388815403,
+ "learning_rate": 0.001247220415690769,
+ "loss": 3.9977524280548096,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.009650378488004208,
+ "learning_rate": 0.001245909441013701,
+ "loss": 3.9882829189300537,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.009350919164717197,
+ "learning_rate": 0.0012445986831380366,
+ "loss": 4.019596099853516,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.008726303465664387,
+ "learning_rate": 0.0012432881431198289,
+ "loss": 3.9933621883392334,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.00829484686255455,
+ "learning_rate": 0.0012419778220149568,
+ "loss": 4.001237869262695,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.008744452148675919,
+ "learning_rate": 0.0012406677208791218,
+ "loss": 3.9323887825012207,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.008170492947101593,
+ "learning_rate": 0.0012393578407678496,
+ "loss": 3.925581455230713,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.007843422703444958,
+ "learning_rate": 0.001238048182736486,
+ "loss": 4.001023769378662,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.008914309553802013,
+ "learning_rate": 0.0012367387478401987,
+ "loss": 3.9804537296295166,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.009736751206219196,
+ "learning_rate": 0.0012354295371339774,
+ "loss": 3.9831767082214355,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.010061381384730339,
+ "learning_rate": 0.0012341205516726273,
+ "loss": 3.9462435245513916,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.010350692085921764,
+ "learning_rate": 0.0012328117925107765,
+ "loss": 4.012863636016846,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01039902213960886,
+ "learning_rate": 0.0012315032607028667,
+ "loss": 3.9811301231384277,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.011423371732234955,
+ "learning_rate": 0.0012301949573031595,
+ "loss": 3.9773542881011963,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.012449614703655243,
+ "learning_rate": 0.001228886883365731,
+ "loss": 3.969600200653076,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.010361429303884506,
+ "learning_rate": 0.0012275790399444726,
+ "loss": 4.035360336303711,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.009603082202374935,
+ "learning_rate": 0.0012262714280930904,
+ "loss": 3.9840846061706543,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.009183845482766628,
+ "learning_rate": 0.0012249640488651036,
+ "loss": 4.031542778015137,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.009694446809589863,
+ "learning_rate": 0.0012236569033138441,
+ "loss": 3.9935057163238525,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.0097816726192832,
+ "learning_rate": 0.0012223499924924545,
+ "loss": 3.9851176738739014,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.009413184598088264,
+ "learning_rate": 0.001221043317453891,
+ "loss": 3.9642205238342285,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.009960711002349854,
+ "learning_rate": 0.0012197368792509166,
+ "loss": 3.9373817443847656,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.010124670341610909,
+ "learning_rate": 0.0012184306789361059,
+ "loss": 3.958115577697754,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.011087656021118164,
+ "learning_rate": 0.00121712471756184,
+ "loss": 3.9607057571411133,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.009812800213694572,
+ "learning_rate": 0.001215818996180309,
+ "loss": 4.00722599029541,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.010305573232471943,
+ "learning_rate": 0.0012145135158435094,
+ "loss": 3.98728084564209,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.01019787136465311,
+ "learning_rate": 0.0012132082776032422,
+ "loss": 3.981172800064087,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.009260181337594986,
+ "learning_rate": 0.001211903282511115,
+ "loss": 4.034339904785156,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.00972998607903719,
+ "learning_rate": 0.001210598531618538,
+ "loss": 4.022885322570801,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.009853586554527283,
+ "learning_rate": 0.0012092940259767267,
+ "loss": 4.000235080718994,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.00961449183523655,
+ "learning_rate": 0.0012079897666366962,
+ "loss": 3.9951353073120117,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.010257176123559475,
+ "learning_rate": 0.0012066857546492657,
+ "loss": 3.9801371097564697,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.009443561546504498,
+ "learning_rate": 0.001205381991065054,
+ "loss": 3.9596502780914307,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.010389018803834915,
+ "learning_rate": 0.0012040784769344795,
+ "loss": 3.980527639389038,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.010322567075490952,
+ "learning_rate": 0.0012027752133077607,
+ "loss": 3.996258497238159,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.009987728670239449,
+ "learning_rate": 0.0012014722012349125,
+ "loss": 3.9493203163146973,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.421086311340332,
+ "eval_runtime": 52.5623,
+ "eval_samples_per_second": 46.459,
+ "eval_steps_per_second": 1.465,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.009405115619301796,
+ "learning_rate": 0.0012001694417657493,
+ "loss": 3.927506923675537,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.009974643588066101,
+ "learning_rate": 0.00119886693594988,
+ "loss": 3.9373550415039062,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.011331096291542053,
+ "learning_rate": 0.0011975646848367098,
+ "loss": 3.935594081878662,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.010538448579609394,
+ "learning_rate": 0.0011962626894754405,
+ "loss": 3.911579132080078,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.00989572424441576,
+ "learning_rate": 0.0011949609509150645,
+ "loss": 3.92525577545166,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.010533696971833706,
+ "learning_rate": 0.0011936594702043704,
+ "loss": 3.898075580596924,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.01051455456763506,
+ "learning_rate": 0.0011923582483919365,
+ "loss": 3.93463134765625,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.009662075899541378,
+ "learning_rate": 0.0011910572865261348,
+ "loss": 3.898221731185913,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.008874096907675266,
+ "learning_rate": 0.0011897565856551257,
+ "loss": 3.926787853240967,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.008126807399094105,
+ "learning_rate": 0.0011884561468268606,
+ "loss": 3.9158413410186768,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.007915553636848927,
+ "learning_rate": 0.001187155971089081,
+ "loss": 3.8891496658325195,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.008312922902405262,
+ "learning_rate": 0.0011858560594893129,
+ "loss": 3.932219982147217,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.008892328478395939,
+ "learning_rate": 0.0011845564130748734,
+ "loss": 3.8866336345672607,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.009187526069581509,
+ "learning_rate": 0.0011832570328928628,
+ "loss": 3.9464707374572754,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.008615494705736637,
+ "learning_rate": 0.0011819579199901693,
+ "loss": 3.9396629333496094,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.008649308234453201,
+ "learning_rate": 0.0011806590754134636,
+ "loss": 3.9325060844421387,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.008581082336604595,
+ "learning_rate": 0.0011793605002092021,
+ "loss": 3.9200944900512695,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.008502699434757233,
+ "learning_rate": 0.0011780621954236238,
+ "loss": 3.9656100273132324,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.0083576999604702,
+ "learning_rate": 0.0011767641621027477,
+ "loss": 3.9130215644836426,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.008196269161999226,
+ "learning_rate": 0.0011754664012923785,
+ "loss": 3.942369222640991,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.008857307955622673,
+ "learning_rate": 0.0011741689140380956,
+ "loss": 3.909585952758789,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.008935349993407726,
+ "learning_rate": 0.0011728717013852635,
+ "loss": 3.9248366355895996,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.008821200579404831,
+ "learning_rate": 0.0011715747643790211,
+ "loss": 3.9589288234710693,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.00861942209303379,
+ "learning_rate": 0.0011702781040642885,
+ "loss": 3.9381728172302246,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.009127574972808361,
+ "learning_rate": 0.0011689817214857613,
+ "loss": 3.897829532623291,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.008008269593119621,
+ "learning_rate": 0.0011676856176879107,
+ "loss": 3.930056095123291,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.010070208460092545,
+ "learning_rate": 0.0011663897937149853,
+ "loss": 3.892270803451538,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.00903113093227148,
+ "learning_rate": 0.001165094250611006,
+ "loss": 3.9766440391540527,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.008872415870428085,
+ "learning_rate": 0.0011637989894197692,
+ "loss": 3.9559712409973145,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.009524816647171974,
+ "learning_rate": 0.0011625040111848427,
+ "loss": 3.9607882499694824,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.009836779907345772,
+ "learning_rate": 0.0011612093169495678,
+ "loss": 3.9629592895507812,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.009586540050804615,
+ "learning_rate": 0.0011599149077570561,
+ "loss": 3.95208740234375,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.010495901107788086,
+ "learning_rate": 0.0011586207846501896,
+ "loss": 3.9439730644226074,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.010346266441047192,
+ "learning_rate": 0.0011573269486716202,
+ "loss": 3.959299087524414,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.010767634026706219,
+ "learning_rate": 0.0011560334008637674,
+ "loss": 3.900784969329834,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.012122200801968575,
+ "learning_rate": 0.0011547401422688202,
+ "loss": 3.9588193893432617,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.012914645485579967,
+ "learning_rate": 0.0011534471739287331,
+ "loss": 3.9278407096862793,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.011785333044826984,
+ "learning_rate": 0.0011521544968852273,
+ "loss": 3.992793560028076,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.008455781266093254,
+ "learning_rate": 0.0011508621121797899,
+ "loss": 3.96525239944458,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.009655282832682133,
+ "learning_rate": 0.0011495700208536714,
+ "loss": 3.9346890449523926,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.010537004098296165,
+ "learning_rate": 0.0011482782239478871,
+ "loss": 3.9534668922424316,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.009438651613891125,
+ "learning_rate": 0.0011469867225032135,
+ "loss": 3.983487606048584,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.010586646385490894,
+ "learning_rate": 0.0011456955175601908,
+ "loss": 3.939680576324463,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.009082718752324581,
+ "learning_rate": 0.001144404610159119,
+ "loss": 3.9558515548706055,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.009833602234721184,
+ "learning_rate": 0.0011431140013400591,
+ "loss": 3.9539828300476074,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.009142170660197735,
+ "learning_rate": 0.0011418236921428313,
+ "loss": 3.9442849159240723,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.008603853173553944,
+ "learning_rate": 0.0011405336836070137,
+ "loss": 3.9526913166046143,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.008838302455842495,
+ "learning_rate": 0.0011392439767719446,
+ "loss": 3.973104953765869,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.008867006748914719,
+ "learning_rate": 0.0011379545726767156,
+ "loss": 3.9832305908203125,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.00971697736531496,
+ "learning_rate": 0.001136665472360178,
+ "loss": 3.9563937187194824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.00957175437361002,
+ "learning_rate": 0.001135376676860935,
+ "loss": 3.975194215774536,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.008548660203814507,
+ "learning_rate": 0.001134088187217347,
+ "loss": 3.9839985370635986,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.008556334301829338,
+ "learning_rate": 0.0011328000044675264,
+ "loss": 3.9225516319274902,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.008440867997705936,
+ "learning_rate": 0.0011315121296493385,
+ "loss": 3.945028066635132,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.009643333032727242,
+ "learning_rate": 0.0011302245638004017,
+ "loss": 3.943120002746582,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.008832835592329502,
+ "learning_rate": 0.001128937307958083,
+ "loss": 3.9895920753479004,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.008720474317669868,
+ "learning_rate": 0.0011276503631595036,
+ "loss": 3.941950798034668,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.008099926635622978,
+ "learning_rate": 0.001126363730441529,
+ "loss": 3.9626870155334473,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.008525107055902481,
+ "learning_rate": 0.0011250774108407778,
+ "loss": 3.950535774230957,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.008998272009193897,
+ "learning_rate": 0.0011237914053936135,
+ "loss": 3.9460678100585938,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.009387437254190445,
+ "learning_rate": 0.001122505715136148,
+ "loss": 3.926604986190796,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.007804017513990402,
+ "learning_rate": 0.0011212203411042386,
+ "loss": 3.9894230365753174,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.008564616553485394,
+ "learning_rate": 0.0011199352843334876,
+ "loss": 3.9339959621429443,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.008726440370082855,
+ "learning_rate": 0.0011186505458592427,
+ "loss": 3.9867987632751465,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.008074100129306316,
+ "learning_rate": 0.0011173661267165932,
+ "loss": 3.9492292404174805,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.007688041776418686,
+ "learning_rate": 0.0011160820279403744,
+ "loss": 3.9425454139709473,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.007808993570506573,
+ "learning_rate": 0.0011147982505651595,
+ "loss": 3.935114860534668,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.009786820039153099,
+ "learning_rate": 0.0011135147956252659,
+ "loss": 3.959758758544922,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.009487155824899673,
+ "learning_rate": 0.00111223166415475,
+ "loss": 3.989048480987549,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.008407804183661938,
+ "learning_rate": 0.0011109488571874073,
+ "loss": 3.932347297668457,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.007158380467444658,
+ "learning_rate": 0.0011096663757567723,
+ "loss": 3.9545154571533203,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.008840581402182579,
+ "learning_rate": 0.001108384220896117,
+ "loss": 3.965984344482422,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.409194469451904,
+ "eval_runtime": 51.5561,
+ "eval_samples_per_second": 47.366,
+ "eval_steps_per_second": 1.494,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.010702896863222122,
+ "learning_rate": 0.0011071023936384508,
+ "loss": 3.8964462280273438,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01290619745850563,
+ "learning_rate": 0.0011058208950165183,
+ "loss": 3.8775625228881836,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01443667896091938,
+ "learning_rate": 0.0011045397260628,
+ "loss": 3.8755288124084473,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.014411968179047108,
+ "learning_rate": 0.0011032588878095108,
+ "loss": 3.889918327331543,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.013656266033649445,
+ "learning_rate": 0.0011019783812885986,
+ "loss": 3.926997661590576,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.010661937296390533,
+ "learning_rate": 0.0011006982075317446,
+ "loss": 3.9226956367492676,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.008968768641352654,
+ "learning_rate": 0.0010994183675703615,
+ "loss": 3.8932647705078125,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.009992181323468685,
+ "learning_rate": 0.0010981388624355937,
+ "loss": 3.8920068740844727,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.011034460738301277,
+ "learning_rate": 0.0010968596931583148,
+ "loss": 3.909073829650879,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.011815821751952171,
+ "learning_rate": 0.0010955808607691284,
+ "loss": 3.9190006256103516,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.012317490763962269,
+ "learning_rate": 0.0010943023662983674,
+ "loss": 3.8790552616119385,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.010983958840370178,
+ "learning_rate": 0.0010930242107760903,
+ "loss": 3.8730852603912354,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.008948802947998047,
+ "learning_rate": 0.0010917463952320856,
+ "loss": 3.923659563064575,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.009568693116307259,
+ "learning_rate": 0.0010904689206958649,
+ "loss": 3.9100794792175293,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.011862963438034058,
+ "learning_rate": 0.0010891917881966672,
+ "loss": 3.908093214035034,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.010180995799601078,
+ "learning_rate": 0.0010879149987634545,
+ "loss": 3.900660514831543,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.009372464381158352,
+ "learning_rate": 0.0010866385534249131,
+ "loss": 3.8950703144073486,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.010695629753172398,
+ "learning_rate": 0.0010853624532094525,
+ "loss": 3.913530111312866,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.011817459017038345,
+ "learning_rate": 0.0010840866991452024,
+ "loss": 3.90268611907959,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.010686337016522884,
+ "learning_rate": 0.0010828112922600167,
+ "loss": 3.9141364097595215,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.009550466202199459,
+ "learning_rate": 0.0010815362335814655,
+ "loss": 3.968468189239502,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.009367208927869797,
+ "learning_rate": 0.0010802615241368423,
+ "loss": 3.8933968544006348,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.009899397380650043,
+ "learning_rate": 0.0010789871649531562,
+ "loss": 3.936708450317383,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.009181005880236626,
+ "learning_rate": 0.0010777131570571363,
+ "loss": 3.9370503425598145,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.00799303874373436,
+ "learning_rate": 0.0010764395014752274,
+ "loss": 3.911050319671631,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.008469315245747566,
+ "learning_rate": 0.0010751661992335908,
+ "loss": 3.9103240966796875,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.009814459830522537,
+ "learning_rate": 0.0010738932513581034,
+ "loss": 3.881843090057373,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.010244108736515045,
+ "learning_rate": 0.0010726206588743557,
+ "loss": 3.883141040802002,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.008898732252418995,
+ "learning_rate": 0.001071348422807653,
+ "loss": 3.952347755432129,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.00823870487511158,
+ "learning_rate": 0.0010700765441830121,
+ "loss": 3.932507038116455,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.009857657365500927,
+ "learning_rate": 0.001068805024025164,
+ "loss": 3.932932138442993,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.00929308496415615,
+ "learning_rate": 0.0010675338633585488,
+ "loss": 3.8971338272094727,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.008667961694300175,
+ "learning_rate": 0.0010662630632073172,
+ "loss": 3.916931629180908,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.009758287109434605,
+ "learning_rate": 0.001064992624595331,
+ "loss": 3.964726209640503,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.009429254569113255,
+ "learning_rate": 0.0010637225485461586,
+ "loss": 3.917572498321533,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.008303353562951088,
+ "learning_rate": 0.0010624528360830782,
+ "loss": 3.875241756439209,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.008729013614356518,
+ "learning_rate": 0.001061183488229073,
+ "loss": 3.9691600799560547,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.010239302180707455,
+ "learning_rate": 0.0010599145060068353,
+ "loss": 3.9053149223327637,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.00932721421122551,
+ "learning_rate": 0.0010586458904387596,
+ "loss": 3.8593955039978027,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.008196988143026829,
+ "learning_rate": 0.0010573776425469473,
+ "loss": 3.917983055114746,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.008049141615629196,
+ "learning_rate": 0.001056109763353203,
+ "loss": 3.9337785243988037,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.009256313554942608,
+ "learning_rate": 0.001054842253879034,
+ "loss": 3.937887668609619,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.009292179718613625,
+ "learning_rate": 0.0010535751151456495,
+ "loss": 3.9203450679779053,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.009142161346971989,
+ "learning_rate": 0.0010523083481739602,
+ "loss": 3.9027457237243652,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.009498288854956627,
+ "learning_rate": 0.0010510419539845777,
+ "loss": 3.917316436767578,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.00849096104502678,
+ "learning_rate": 0.001049775933597813,
+ "loss": 3.9229235649108887,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.008343606255948544,
+ "learning_rate": 0.0010485102880336747,
+ "loss": 3.9138073921203613,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.010342134162783623,
+ "learning_rate": 0.0010472450183118727,
+ "loss": 3.8932929039001465,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.010298306122422218,
+ "learning_rate": 0.00104598012545181,
+ "loss": 3.9112677574157715,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.00827560480684042,
+ "learning_rate": 0.001044715610472589,
+ "loss": 3.894556760787964,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.008993102237582207,
+ "learning_rate": 0.0010434514743930064,
+ "loss": 3.905391216278076,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.008632538840174675,
+ "learning_rate": 0.0010421877182315536,
+ "loss": 3.896714210510254,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.008542424067854881,
+ "learning_rate": 0.0010409243430064162,
+ "loss": 3.9145569801330566,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.008970450609922409,
+ "learning_rate": 0.0010396613497354725,
+ "loss": 3.9170360565185547,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.008744142949581146,
+ "learning_rate": 0.0010383987394362942,
+ "loss": 3.927412509918213,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.008229847066104412,
+ "learning_rate": 0.0010371365131261424,
+ "loss": 3.928352117538452,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.00818693358451128,
+ "learning_rate": 0.0010358746718219717,
+ "loss": 3.9219343662261963,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.008848588913679123,
+ "learning_rate": 0.001034613216540423,
+ "loss": 3.9346699714660645,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.00834780652076006,
+ "learning_rate": 0.0010333521482978295,
+ "loss": 3.9144492149353027,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.00940527394413948,
+ "learning_rate": 0.0010320914681102098,
+ "loss": 3.951573371887207,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.008884521201252937,
+ "learning_rate": 0.0010308311769932725,
+ "loss": 3.899625778198242,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.007863261736929417,
+ "learning_rate": 0.0010295712759624107,
+ "loss": 3.9321727752685547,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.008651153184473515,
+ "learning_rate": 0.0010283117660327035,
+ "loss": 3.9030094146728516,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.00822074618190527,
+ "learning_rate": 0.0010270526482189163,
+ "loss": 3.939466953277588,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.008701153099536896,
+ "learning_rate": 0.0010257939235354964,
+ "loss": 3.9099626541137695,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.008418620564043522,
+ "learning_rate": 0.0010245355929965774,
+ "loss": 3.914116382598877,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.008339700289070606,
+ "learning_rate": 0.001023277657615971,
+ "loss": 3.9111878871917725,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.008304647170007229,
+ "learning_rate": 0.0010220201184071747,
+ "loss": 3.950974464416504,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.008436229079961777,
+ "learning_rate": 0.001020762976383365,
+ "loss": 3.9631457328796387,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.0074487063102424145,
+ "learning_rate": 0.001019506232557398,
+ "loss": 3.902984619140625,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.008736515417695045,
+ "learning_rate": 0.0010182498879418096,
+ "loss": 3.9268949031829834,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.009666468016803265,
+ "learning_rate": 0.0010169939435488145,
+ "loss": 3.9274301528930664,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.395970821380615,
+ "eval_runtime": 53.7848,
+ "eval_samples_per_second": 45.403,
+ "eval_steps_per_second": 1.432,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.009094545617699623,
+ "learning_rate": 0.001015738400390304,
+ "loss": 3.866425037384033,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.009273733012378216,
+ "learning_rate": 0.0010144832594778462,
+ "loss": 3.8569841384887695,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.009205271489918232,
+ "learning_rate": 0.0010132285218226855,
+ "loss": 3.8981378078460693,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.009963045828044415,
+ "learning_rate": 0.0010119741884357424,
+ "loss": 3.881542444229126,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.008536554872989655,
+ "learning_rate": 0.0010107202603276096,
+ "loss": 3.851253032684326,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.01074561569839716,
+ "learning_rate": 0.001009466738508555,
+ "loss": 3.8731184005737305,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.010809902101755142,
+ "learning_rate": 0.0010082136239885177,
+ "loss": 3.863889217376709,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.008289352990686893,
+ "learning_rate": 0.0010069609177771103,
+ "loss": 3.839066743850708,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.00792208593338728,
+ "learning_rate": 0.001005708620883615,
+ "loss": 3.8634843826293945,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.008381598629057407,
+ "learning_rate": 0.0010044567343169845,
+ "loss": 3.868283987045288,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.009784617461264133,
+ "learning_rate": 0.0010032052590858425,
+ "loss": 3.853074550628662,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.009462198242545128,
+ "learning_rate": 0.0010019541961984782,
+ "loss": 3.889979600906372,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.010207819752395153,
+ "learning_rate": 0.001000703546662852,
+ "loss": 3.8636741638183594,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.009123551659286022,
+ "learning_rate": 0.0009994533114865887,
+ "loss": 3.8578555583953857,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.009006140753626823,
+ "learning_rate": 0.0009982034916769806,
+ "loss": 3.8828015327453613,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.0091114342212677,
+ "learning_rate": 0.0009969540882409846,
+ "loss": 3.880486011505127,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.008851367980241776,
+ "learning_rate": 0.0009957051021852227,
+ "loss": 3.8618125915527344,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.008196741342544556,
+ "learning_rate": 0.0009944565345159806,
+ "loss": 3.8314175605773926,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.00863513071089983,
+ "learning_rate": 0.000993208386239206,
+ "loss": 3.856212615966797,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.008624681271612644,
+ "learning_rate": 0.0009919606583605107,
+ "loss": 3.868927001953125,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.009208842180669308,
+ "learning_rate": 0.0009907133518851647,
+ "loss": 3.9072189331054688,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.010200913064181805,
+ "learning_rate": 0.0009894664678181027,
+ "loss": 3.875272750854492,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.009892710484564304,
+ "learning_rate": 0.000988220007163914,
+ "loss": 3.888686418533325,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.009873585775494576,
+ "learning_rate": 0.000986973970926851,
+ "loss": 3.8607118129730225,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.009322221390902996,
+ "learning_rate": 0.0009857283601108228,
+ "loss": 3.8852052688598633,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.008141528815031052,
+ "learning_rate": 0.0009844831757193947,
+ "loss": 3.888213872909546,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.008191613480448723,
+ "learning_rate": 0.00098323841875579,
+ "loss": 3.9073967933654785,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.008419081568717957,
+ "learning_rate": 0.0009819940902228864,
+ "loss": 3.879223346710205,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.009037869982421398,
+ "learning_rate": 0.0009807501911232171,
+ "loss": 3.883303165435791,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.00854078121483326,
+ "learning_rate": 0.0009795067224589687,
+ "loss": 3.926767587661743,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.008590771816670895,
+ "learning_rate": 0.0009782636852319825,
+ "loss": 3.913999557495117,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.00859815627336502,
+ "learning_rate": 0.0009770210804437498,
+ "loss": 3.8828063011169434,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.008970260620117188,
+ "learning_rate": 0.0009757789090954157,
+ "loss": 3.9174866676330566,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.00877175759524107,
+ "learning_rate": 0.0009745371721877754,
+ "loss": 3.8879401683807373,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.008541764691472054,
+ "learning_rate": 0.0009732958707212731,
+ "loss": 3.8717427253723145,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.008466241881251335,
+ "learning_rate": 0.0009720550056960039,
+ "loss": 3.8944485187530518,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.00808491837233305,
+ "learning_rate": 0.0009708145781117096,
+ "loss": 3.8617026805877686,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.00842207670211792,
+ "learning_rate": 0.0009695745889677805,
+ "loss": 3.887824535369873,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.00830599945038557,
+ "learning_rate": 0.0009683350392632533,
+ "loss": 3.8612070083618164,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.008417914621531963,
+ "learning_rate": 0.0009670959299968111,
+ "loss": 3.880140781402588,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.008542059920728207,
+ "learning_rate": 0.0009658572621667822,
+ "loss": 3.8951029777526855,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.008822236210107803,
+ "learning_rate": 0.000964619036771138,
+ "loss": 3.8962583541870117,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.009151184931397438,
+ "learning_rate": 0.0009633812548074952,
+ "loss": 3.882500648498535,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.007803036365658045,
+ "learning_rate": 0.0009621439172731115,
+ "loss": 3.8740782737731934,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.009350850246846676,
+ "learning_rate": 0.0009609070251648884,
+ "loss": 3.8627419471740723,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.008319144137203693,
+ "learning_rate": 0.0009596705794793667,
+ "loss": 3.9084486961364746,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.009198528714478016,
+ "learning_rate": 0.0009584345812127282,
+ "loss": 3.92141056060791,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.010480938479304314,
+ "learning_rate": 0.0009571990313607958,
+ "loss": 3.8916103839874268,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.01045288611203432,
+ "learning_rate": 0.0009559639309190278,
+ "loss": 3.8863210678100586,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.009121391922235489,
+ "learning_rate": 0.0009547292808825239,
+ "loss": 3.9167044162750244,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.01093223039060831,
+ "learning_rate": 0.0009534950822460184,
+ "loss": 3.9046711921691895,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.01214868575334549,
+ "learning_rate": 0.0009522613360038835,
+ "loss": 3.9126768112182617,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.01055280864238739,
+ "learning_rate": 0.0009510280431501255,
+ "loss": 3.9214353561401367,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.009021313861012459,
+ "learning_rate": 0.0009497952046783867,
+ "loss": 3.8795580863952637,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.00863752793520689,
+ "learning_rate": 0.0009485628215819428,
+ "loss": 3.893232822418213,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.009327626787126064,
+ "learning_rate": 0.0009473308948537021,
+ "loss": 3.917377471923828,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.01003898587077856,
+ "learning_rate": 0.0009460994254862066,
+ "loss": 3.8818931579589844,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.00892532430589199,
+ "learning_rate": 0.0009448684144716278,
+ "loss": 3.8761978149414062,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.008396463468670845,
+ "learning_rate": 0.0009436378628017702,
+ "loss": 3.8665928840637207,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.00784370768815279,
+ "learning_rate": 0.0009424077714680662,
+ "loss": 3.8850138187408447,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.00809321366250515,
+ "learning_rate": 0.0009411781414615784,
+ "loss": 3.917078733444214,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.008685288950800896,
+ "learning_rate": 0.000939948973772998,
+ "loss": 3.8865013122558594,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.007291820365935564,
+ "learning_rate": 0.0009387202693926425,
+ "loss": 3.8967666625976562,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.008459319360554218,
+ "learning_rate": 0.0009374920293104573,
+ "loss": 3.9043235778808594,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.007924137637019157,
+ "learning_rate": 0.0009362642545160123,
+ "loss": 3.9216856956481934,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.007874754257500172,
+ "learning_rate": 0.0009350369459985052,
+ "loss": 3.879725933074951,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.008643723092973232,
+ "learning_rate": 0.0009338101047467544,
+ "loss": 3.9303886890411377,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.008451704867184162,
+ "learning_rate": 0.000932583731749205,
+ "loss": 3.9352827072143555,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.008137754164636135,
+ "learning_rate": 0.0009313578279939236,
+ "loss": 3.930800437927246,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.009277567267417908,
+ "learning_rate": 0.0009301323944685979,
+ "loss": 3.9194300174713135,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.009428970515727997,
+ "learning_rate": 0.0009289074321605383,
+ "loss": 3.874298572540283,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.0089727733284235,
+ "learning_rate": 0.0009276829420566743,
+ "loss": 3.893770694732666,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.388528823852539,
+ "eval_runtime": 51.5859,
+ "eval_samples_per_second": 47.338,
+ "eval_steps_per_second": 1.493,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.009470812045037746,
+ "learning_rate": 0.000926458925143556,
+ "loss": 3.8514318466186523,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.011485518887639046,
+ "learning_rate": 0.0009252353824073511,
+ "loss": 3.857734203338623,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.010095939971506596,
+ "learning_rate": 0.0009240123148338457,
+ "loss": 3.819117546081543,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.010103728622198105,
+ "learning_rate": 0.0009227897234084444,
+ "loss": 3.842660427093506,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.01148147787898779,
+ "learning_rate": 0.0009215676091161661,
+ "loss": 3.821089029312134,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.012331553734838963,
+ "learning_rate": 0.0009203459729416467,
+ "loss": 3.8194339275360107,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.011274794116616249,
+ "learning_rate": 0.0009191248158691358,
+ "loss": 3.8839240074157715,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.009603365324437618,
+ "learning_rate": 0.0009179041388824986,
+ "loss": 3.806617259979248,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.010824795812368393,
+ "learning_rate": 0.0009166839429652112,
+ "loss": 3.823228597640991,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.01014280878007412,
+ "learning_rate": 0.0009154642291003637,
+ "loss": 3.860605239868164,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.009052124805748463,
+ "learning_rate": 0.0009142449982706591,
+ "loss": 3.84226131439209,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.009465130977332592,
+ "learning_rate": 0.0009130262514584074,
+ "loss": 3.8674912452697754,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.009991783648729324,
+ "learning_rate": 0.0009118079896455326,
+ "loss": 3.8775739669799805,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.009288253262639046,
+ "learning_rate": 0.0009105902138135652,
+ "loss": 3.7978098392486572,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.009964323602616787,
+ "learning_rate": 0.0009093729249436459,
+ "loss": 3.8362321853637695,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.010977157391607761,
+ "learning_rate": 0.0009081561240165218,
+ "loss": 3.857304096221924,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.009241319261491299,
+ "learning_rate": 0.0009069398120125479,
+ "loss": 3.8659310340881348,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.009211038239300251,
+ "learning_rate": 0.0009057239899116844,
+ "loss": 3.8764538764953613,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.009649050422012806,
+ "learning_rate": 0.000904508658693497,
+ "loss": 3.8702778816223145,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.00954380165785551,
+ "learning_rate": 0.0009032938193371573,
+ "loss": 3.8273327350616455,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.008831270970404148,
+ "learning_rate": 0.0009020794728214378,
+ "loss": 3.821664810180664,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.008367286063730717,
+ "learning_rate": 0.000900865620124717,
+ "loss": 3.8586034774780273,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.008805174380540848,
+ "learning_rate": 0.0008996522622249732,
+ "loss": 3.811522960662842,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.00873226672410965,
+ "learning_rate": 0.0008984394000997873,
+ "loss": 3.8538150787353516,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.008930203504860401,
+ "learning_rate": 0.0008972270347263402,
+ "loss": 3.8448486328125,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.009300434030592442,
+ "learning_rate": 0.0008960151670814125,
+ "loss": 3.842113494873047,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.009598663076758385,
+ "learning_rate": 0.0008948037981413855,
+ "loss": 3.886852741241455,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.008400370366871357,
+ "learning_rate": 0.0008935929288822357,
+ "loss": 3.852076292037964,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.00891344528645277,
+ "learning_rate": 0.0008923825602795391,
+ "loss": 3.8480606079101562,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.00869396049529314,
+ "learning_rate": 0.0008911726933084678,
+ "loss": 3.8164706230163574,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.007914266549050808,
+ "learning_rate": 0.0008899633289437904,
+ "loss": 3.8768272399902344,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.008398533798754215,
+ "learning_rate": 0.0008887544681598687,
+ "loss": 3.8625502586364746,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.00887146033346653,
+ "learning_rate": 0.0008875461119306601,
+ "loss": 3.8934593200683594,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.008642833679914474,
+ "learning_rate": 0.0008863382612297172,
+ "loss": 3.8391976356506348,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.008099788799881935,
+ "learning_rate": 0.0008851309170301816,
+ "loss": 3.827965497970581,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.008166342042386532,
+ "learning_rate": 0.0008839240803047894,
+ "loss": 3.8999152183532715,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.009095479734241962,
+ "learning_rate": 0.0008827177520258671,
+ "loss": 3.876927614212036,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.009246554225683212,
+ "learning_rate": 0.0008815119331653321,
+ "loss": 3.8692450523376465,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.007059223018586636,
+ "learning_rate": 0.0008803066246946898,
+ "loss": 3.869400978088379,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.008499744348227978,
+ "learning_rate": 0.0008791018275850363,
+ "loss": 3.8303189277648926,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.007667839527130127,
+ "learning_rate": 0.0008778975428070543,
+ "loss": 3.827998638153076,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.00765089550986886,
+ "learning_rate": 0.0008766937713310145,
+ "loss": 3.8554482460021973,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.00862938817590475,
+ "learning_rate": 0.0008754905141267749,
+ "loss": 3.905458927154541,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.008733712136745453,
+ "learning_rate": 0.0008742877721637758,
+ "loss": 3.8564133644104004,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.0071738120168447495,
+ "learning_rate": 0.0008730855464110471,
+ "loss": 3.878807544708252,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.008418241515755653,
+ "learning_rate": 0.0008718838378371989,
+ "loss": 3.8399720191955566,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.008526502177119255,
+ "learning_rate": 0.0008706826474104263,
+ "loss": 3.847656488418579,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.00732462527230382,
+ "learning_rate": 0.0008694819760985069,
+ "loss": 3.8839492797851562,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.008424196392297745,
+ "learning_rate": 0.0008682818248688003,
+ "loss": 3.861239433288574,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.008984955959022045,
+ "learning_rate": 0.0008670821946882465,
+ "loss": 3.856529712677002,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.009004034101963043,
+ "learning_rate": 0.000865883086523365,
+ "loss": 3.8514528274536133,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.009097260423004627,
+ "learning_rate": 0.0008646845013402563,
+ "loss": 3.8890581130981445,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.009244748391211033,
+ "learning_rate": 0.0008634864401045988,
+ "loss": 3.872745990753174,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.00788815412670374,
+ "learning_rate": 0.0008622889037816488,
+ "loss": 3.8777575492858887,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.008943610824644566,
+ "learning_rate": 0.0008610918933362396,
+ "loss": 3.8825736045837402,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.008154108189046383,
+ "learning_rate": 0.0008598954097327809,
+ "loss": 3.877514362335205,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.008689315989613533,
+ "learning_rate": 0.0008586994539352587,
+ "loss": 3.854720115661621,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.00904505793005228,
+ "learning_rate": 0.0008575040269072315,
+ "loss": 3.911463499069214,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.007773210760205984,
+ "learning_rate": 0.0008563091296118342,
+ "loss": 3.875370740890503,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.008437697775661945,
+ "learning_rate": 0.0008551147630117737,
+ "loss": 3.8457469940185547,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.009780957363545895,
+ "learning_rate": 0.0008539209280693296,
+ "loss": 3.890479564666748,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.007929213345050812,
+ "learning_rate": 0.0008527276257463532,
+ "loss": 3.876030445098877,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.009121386334300041,
+ "learning_rate": 0.0008515348570042669,
+ "loss": 3.9252734184265137,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.00981531199067831,
+ "learning_rate": 0.000850342622804063,
+ "loss": 3.8997983932495117,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.009150844067335129,
+ "learning_rate": 0.0008491509241063025,
+ "loss": 3.85115385055542,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.00992978923022747,
+ "learning_rate": 0.0008479597618711154,
+ "loss": 3.9138102531433105,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.010906689800322056,
+ "learning_rate": 0.0008467691370582001,
+ "loss": 3.892329692840576,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.008644519373774529,
+ "learning_rate": 0.0008455790506268213,
+ "loss": 3.9071311950683594,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.009030954912304878,
+ "learning_rate": 0.0008443895035358111,
+ "loss": 3.867384910583496,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.010806960985064507,
+ "learning_rate": 0.0008432004967435638,
+ "loss": 3.877983331680298,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.011148087680339813,
+ "learning_rate": 0.0008420120312080436,
+ "loss": 3.8795924186706543,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01046178862452507,
+ "learning_rate": 0.0008408241078867731,
+ "loss": 3.887709617614746,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.383835792541504,
+ "eval_runtime": 53.0778,
+ "eval_samples_per_second": 46.008,
+ "eval_steps_per_second": 1.451,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.009721867740154266,
+ "learning_rate": 0.0008396367277368421,
+ "loss": 3.8261327743530273,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.01091420091688633,
+ "learning_rate": 0.000838449891714901,
+ "loss": 3.7837276458740234,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.010423852130770683,
+ "learning_rate": 0.0008372636007771623,
+ "loss": 3.796450138092041,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.011928882449865341,
+ "learning_rate": 0.0008360778558793996,
+ "loss": 3.794717311859131,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.010779444128274918,
+ "learning_rate": 0.000834892657976945,
+ "loss": 3.8477983474731445,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.009554282762110233,
+ "learning_rate": 0.0008337080080246915,
+ "loss": 3.793790817260742,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.009768152609467506,
+ "learning_rate": 0.0008325239069770906,
+ "loss": 3.860384225845337,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.010801832191646099,
+ "learning_rate": 0.0008313403557881513,
+ "loss": 3.816796064376831,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.01046004518866539,
+ "learning_rate": 0.0008301573554114378,
+ "loss": 3.8064301013946533,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.009319447912275791,
+ "learning_rate": 0.0008289749068000744,
+ "loss": 3.832408905029297,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01011407095938921,
+ "learning_rate": 0.000827793010906737,
+ "loss": 3.8276009559631348,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.010393989272415638,
+ "learning_rate": 0.0008266116686836582,
+ "loss": 3.83754825592041,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.00883004255592823,
+ "learning_rate": 0.0008254308810826244,
+ "loss": 3.8315513134002686,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.008549787104129791,
+ "learning_rate": 0.0008242506490549745,
+ "loss": 3.792919158935547,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.009668390266597271,
+ "learning_rate": 0.0008230709735516006,
+ "loss": 3.8110954761505127,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.009008699096739292,
+ "learning_rate": 0.0008218918555229445,
+ "loss": 3.8598508834838867,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.008660725317895412,
+ "learning_rate": 0.0008207132959190026,
+ "loss": 3.839632511138916,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.00827972125262022,
+ "learning_rate": 0.0008195352956893172,
+ "loss": 3.8428683280944824,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.008117233403027058,
+ "learning_rate": 0.0008183578557829825,
+ "loss": 3.8434956073760986,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.008314455859363079,
+ "learning_rate": 0.0008171809771486405,
+ "loss": 3.8079686164855957,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.00865850131958723,
+ "learning_rate": 0.0008160046607344811,
+ "loss": 3.846999406814575,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.0096864253282547,
+ "learning_rate": 0.0008148289074882416,
+ "loss": 3.8227243423461914,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.00840236246585846,
+ "learning_rate": 0.0008136537183572045,
+ "loss": 3.8400344848632812,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.007451791316270828,
+ "learning_rate": 0.0008124790942881984,
+ "loss": 3.8318095207214355,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.00858070608228445,
+ "learning_rate": 0.000811305036227597,
+ "loss": 3.8232922554016113,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.00765061192214489,
+ "learning_rate": 0.0008101315451213178,
+ "loss": 3.81624174118042,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.009304090403020382,
+ "learning_rate": 0.0008089586219148213,
+ "loss": 3.8715004920959473,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.009136030450463295,
+ "learning_rate": 0.0008077862675531103,
+ "loss": 3.825037956237793,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.007460939697921276,
+ "learning_rate": 0.0008066144829807303,
+ "loss": 3.811749219894409,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.007510480936616659,
+ "learning_rate": 0.0008054432691417659,
+ "loss": 3.8313679695129395,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.008825462311506271,
+ "learning_rate": 0.0008042726269798433,
+ "loss": 3.83119535446167,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.009002318605780602,
+ "learning_rate": 0.0008031025574381281,
+ "loss": 3.8542468547821045,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.009266122244298458,
+ "learning_rate": 0.0008019330614593237,
+ "loss": 3.8328065872192383,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.008935115300118923,
+ "learning_rate": 0.0008007641399856727,
+ "loss": 3.8198719024658203,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.008021850138902664,
+ "learning_rate": 0.0007995957939589526,
+ "loss": 3.828382968902588,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.008035171777009964,
+ "learning_rate": 0.0007984280243204805,
+ "loss": 3.879868507385254,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.008553579449653625,
+ "learning_rate": 0.0007972608320111059,
+ "loss": 3.842162609100342,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.008016746491193771,
+ "learning_rate": 0.000796094217971215,
+ "loss": 3.85674786567688,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.007781418971717358,
+ "learning_rate": 0.0007949281831407276,
+ "loss": 3.7968859672546387,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.008552255108952522,
+ "learning_rate": 0.0007937627284590971,
+ "loss": 3.8305506706237793,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.007390565704554319,
+ "learning_rate": 0.0007925978548653096,
+ "loss": 3.8228843212127686,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.007611850742250681,
+ "learning_rate": 0.0007914335632978812,
+ "loss": 3.868957996368408,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.008188549429178238,
+ "learning_rate": 0.0007902698546948628,
+ "loss": 3.836078405380249,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.008198797702789307,
+ "learning_rate": 0.0007891067299938318,
+ "loss": 3.8637044429779053,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.008353965356945992,
+ "learning_rate": 0.0007879441901318977,
+ "loss": 3.833843946456909,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.0086967209354043,
+ "learning_rate": 0.0007867822360456963,
+ "loss": 3.850536823272705,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.009151408448815346,
+ "learning_rate": 0.0007856208686713949,
+ "loss": 3.8453407287597656,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.008010942488908768,
+ "learning_rate": 0.0007844600889446858,
+ "loss": 3.8679513931274414,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.007813206873834133,
+ "learning_rate": 0.0007832998978007873,
+ "loss": 3.860609531402588,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.007759213913232088,
+ "learning_rate": 0.0007821402961744448,
+ "loss": 3.803863048553467,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.008045864291489124,
+ "learning_rate": 0.0007809812849999293,
+ "loss": 3.858161449432373,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.008212368935346603,
+ "learning_rate": 0.0007798228652110347,
+ "loss": 3.8311336040496826,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.007982932962477207,
+ "learning_rate": 0.0007786650377410774,
+ "loss": 3.870339870452881,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.0071402136236429214,
+ "learning_rate": 0.0007775078035229002,
+ "loss": 3.7922263145446777,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.008564750663936138,
+ "learning_rate": 0.0007763511634888656,
+ "loss": 3.83981990814209,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.008078585378825665,
+ "learning_rate": 0.0007751951185708562,
+ "loss": 3.852938652038574,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.00801754929125309,
+ "learning_rate": 0.0007740396697002774,
+ "loss": 3.8434009552001953,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.008973612450063229,
+ "learning_rate": 0.0007728848178080535,
+ "loss": 3.8238563537597656,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.010199865326285362,
+ "learning_rate": 0.0007717305638246281,
+ "loss": 3.8891735076904297,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.009620471857488155,
+ "learning_rate": 0.0007705769086799617,
+ "loss": 3.8611769676208496,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.007981563918292522,
+ "learning_rate": 0.0007694238533035337,
+ "loss": 3.8496928215026855,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.008856567554175854,
+ "learning_rate": 0.0007682713986243415,
+ "loss": 3.8366098403930664,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.008323253132402897,
+ "learning_rate": 0.0007671195455708952,
+ "loss": 3.8224244117736816,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.00814893003553152,
+ "learning_rate": 0.0007659682950712228,
+ "loss": 3.8611154556274414,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.009146672673523426,
+ "learning_rate": 0.0007648176480528659,
+ "loss": 3.8258843421936035,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.007573423441499472,
+ "learning_rate": 0.0007636676054428805,
+ "loss": 3.888007640838623,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.00813984964042902,
+ "learning_rate": 0.0007625181681678343,
+ "loss": 3.860306739807129,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.007936778478324413,
+ "learning_rate": 0.0007613693371538078,
+ "loss": 3.8808555603027344,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.008434669114649296,
+ "learning_rate": 0.0007602211133263956,
+ "loss": 3.854154348373413,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.007857564836740494,
+ "learning_rate": 0.0007590734976106985,
+ "loss": 3.8240737915039062,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.007989327423274517,
+ "learning_rate": 0.0007579264909313315,
+ "loss": 3.850971221923828,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.008888393640518188,
+ "learning_rate": 0.0007567800942124152,
+ "loss": 3.862267255783081,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.38022518157959,
+ "eval_runtime": 51.4686,
+ "eval_samples_per_second": 47.446,
+ "eval_steps_per_second": 1.496,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.00952722318470478,
+ "learning_rate": 0.0007556343083775832,
+ "loss": 3.7705495357513428,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01099323108792305,
+ "learning_rate": 0.0007544891343499723,
+ "loss": 3.80812406539917,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.010298571549355984,
+ "learning_rate": 0.0007533445730522295,
+ "loss": 3.782794237136841,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.008139589801430702,
+ "learning_rate": 0.000752200625406507,
+ "loss": 3.7877392768859863,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.009031252935528755,
+ "learning_rate": 0.0007510572923344628,
+ "loss": 3.8181564807891846,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.00942459236830473,
+ "learning_rate": 0.0007499145747572603,
+ "loss": 3.807425022125244,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.00922947097569704,
+ "learning_rate": 0.0007487724735955648,
+ "loss": 3.7923264503479004,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.009551790542900562,
+ "learning_rate": 0.0007476309897695485,
+ "loss": 3.790083885192871,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.008996761403977871,
+ "learning_rate": 0.0007464901241988831,
+ "loss": 3.836491107940674,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.009766182862222195,
+ "learning_rate": 0.0007453498778027433,
+ "loss": 3.7829647064208984,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.007708575110882521,
+ "learning_rate": 0.0007442102514998058,
+ "loss": 3.803400993347168,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.008606902323663235,
+ "learning_rate": 0.0007430712462082464,
+ "loss": 3.810825824737549,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.009462927468121052,
+ "learning_rate": 0.0007419328628457416,
+ "loss": 3.784748077392578,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.010283366777002811,
+ "learning_rate": 0.0007407951023294655,
+ "loss": 3.744312286376953,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.008952794596552849,
+ "learning_rate": 0.0007396579655760917,
+ "loss": 3.8098015785217285,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.008446207270026207,
+ "learning_rate": 0.0007385214535017907,
+ "loss": 3.8306760787963867,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.007988285273313522,
+ "learning_rate": 0.0007373855670222302,
+ "loss": 3.7976696491241455,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.007586437277495861,
+ "learning_rate": 0.0007362503070525718,
+ "loss": 3.785891056060791,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.00775605533272028,
+ "learning_rate": 0.0007351156745074757,
+ "loss": 3.8176379203796387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.007727691903710365,
+ "learning_rate": 0.0007339816703010948,
+ "loss": 3.791743278503418,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.008311517536640167,
+ "learning_rate": 0.0007328482953470747,
+ "loss": 3.818769693374634,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.007610874716192484,
+ "learning_rate": 0.0007317155505585556,
+ "loss": 3.76772141456604,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.0078107016161084175,
+ "learning_rate": 0.0007305834368481699,
+ "loss": 3.822758197784424,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.006906864698976278,
+ "learning_rate": 0.0007294519551280412,
+ "loss": 3.832834005355835,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.008230660110712051,
+ "learning_rate": 0.0007283211063097829,
+ "loss": 3.85087513923645,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.008201118558645248,
+ "learning_rate": 0.0007271908913044997,
+ "loss": 3.7777247428894043,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.0070146857760846615,
+ "learning_rate": 0.000726061311022787,
+ "loss": 3.837517738342285,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.0076036700047552586,
+ "learning_rate": 0.0007249323663747254,
+ "loss": 3.8358592987060547,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.007238952908664942,
+ "learning_rate": 0.0007238040582698858,
+ "loss": 3.828089952468872,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.00825269054621458,
+ "learning_rate": 0.0007226763876173259,
+ "loss": 3.806016445159912,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.008368568494915962,
+ "learning_rate": 0.0007215493553255898,
+ "loss": 3.8163394927978516,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.00728471577167511,
+ "learning_rate": 0.0007204229623027063,
+ "loss": 3.835597038269043,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.00784830842167139,
+ "learning_rate": 0.0007192972094561897,
+ "loss": 3.8050971031188965,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.007714141625910997,
+ "learning_rate": 0.0007181720976930404,
+ "loss": 3.8031606674194336,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.009128103032708168,
+ "learning_rate": 0.000717047627919739,
+ "loss": 3.77107572555542,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.008033272810280323,
+ "learning_rate": 0.000715923801042251,
+ "loss": 3.8615059852600098,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.007732565049082041,
+ "learning_rate": 0.0007148006179660235,
+ "loss": 3.8277132511138916,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.007382966112345457,
+ "learning_rate": 0.0007136780795959853,
+ "loss": 3.803431987762451,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.007517020218074322,
+ "learning_rate": 0.0007125561868365442,
+ "loss": 3.799734354019165,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.008041434921324253,
+ "learning_rate": 0.0007114349405915895,
+ "loss": 3.7848258018493652,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.007875649258494377,
+ "learning_rate": 0.0007103143417644891,
+ "loss": 3.8136563301086426,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.008213640190660954,
+ "learning_rate": 0.000709194391258089,
+ "loss": 3.8518495559692383,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.00801054760813713,
+ "learning_rate": 0.0007080750899747139,
+ "loss": 3.7956271171569824,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.008064007386565208,
+ "learning_rate": 0.0007069564388161631,
+ "loss": 3.854595184326172,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.008300521411001682,
+ "learning_rate": 0.0007058384386837154,
+ "loss": 3.8203625679016113,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.008640496991574764,
+ "learning_rate": 0.000704721090478122,
+ "loss": 3.793574810028076,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.007684200536459684,
+ "learning_rate": 0.0007036043950996105,
+ "loss": 3.8564438819885254,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.008431905880570412,
+ "learning_rate": 0.0007024883534478823,
+ "loss": 3.8271119594573975,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.008137186989188194,
+ "learning_rate": 0.0007013729664221122,
+ "loss": 3.8219337463378906,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.0076744635589420795,
+ "learning_rate": 0.0007002582349209474,
+ "loss": 3.8353049755096436,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.008388740941882133,
+ "learning_rate": 0.0006991441598425065,
+ "loss": 3.795104503631592,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.007856911979615688,
+ "learning_rate": 0.0006980307420843798,
+ "loss": 3.818411111831665,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.008048887364566326,
+ "learning_rate": 0.000696917982543628,
+ "loss": 3.86599063873291,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.007949979975819588,
+ "learning_rate": 0.0006958058821167812,
+ "loss": 3.8026742935180664,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.008162821643054485,
+ "learning_rate": 0.0006946944416998389,
+ "loss": 3.833751678466797,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.008190765976905823,
+ "learning_rate": 0.0006935836621882682,
+ "loss": 3.8164236545562744,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.009714310988783836,
+ "learning_rate": 0.000692473544477005,
+ "loss": 3.8367366790771484,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.008075987920165062,
+ "learning_rate": 0.0006913640894604498,
+ "loss": 3.8108773231506348,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.009075167588889599,
+ "learning_rate": 0.0006902552980324713,
+ "loss": 3.8444485664367676,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.008678440935909748,
+ "learning_rate": 0.0006891471710864022,
+ "loss": 3.8384461402893066,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.007705411408096552,
+ "learning_rate": 0.0006880397095150414,
+ "loss": 3.825443744659424,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.00782099086791277,
+ "learning_rate": 0.0006869329142106488,
+ "loss": 3.8331687450408936,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.007542501203715801,
+ "learning_rate": 0.0006858267860649509,
+ "loss": 3.7905991077423096,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.008315361104905605,
+ "learning_rate": 0.0006847213259691356,
+ "loss": 3.861527442932129,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.008529976941645145,
+ "learning_rate": 0.0006836165348138508,
+ "loss": 3.8332972526550293,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.008117211051285267,
+ "learning_rate": 0.0006825124134892072,
+ "loss": 3.7961089611053467,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.008409497328102589,
+ "learning_rate": 0.0006814089628847755,
+ "loss": 3.812218189239502,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.007592010777443647,
+ "learning_rate": 0.0006803061838895864,
+ "loss": 3.801304340362549,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.007820053026080132,
+ "learning_rate": 0.0006792040773921283,
+ "loss": 3.8517117500305176,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.0073044621385633945,
+ "learning_rate": 0.0006781026442803479,
+ "loss": 3.8253560066223145,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.007546546868979931,
+ "learning_rate": 0.0006770018854416522,
+ "loss": 3.800802230834961,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.008878474123775959,
+ "learning_rate": 0.0006759018017629007,
+ "loss": 3.8133859634399414,
+ "step": 2520
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.0743646764662784e+18,
+ "train_batch_size": 16,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/baseline/checkpoint-2520/training_args.bin b/runs/baseline/checkpoint-2520/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7
--- /dev/null
+++ b/runs/baseline/checkpoint-2520/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97
+size 5265
diff --git a/runs/baseline/checkpoint-2880/chat_template.jinja b/runs/baseline/checkpoint-2880/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/baseline/checkpoint-2880/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/baseline/checkpoint-2880/config.json b/runs/baseline/checkpoint-2880/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/baseline/checkpoint-2880/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/baseline/checkpoint-2880/generation_config.json b/runs/baseline/checkpoint-2880/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/baseline/checkpoint-2880/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/baseline/checkpoint-2880/model.safetensors b/runs/baseline/checkpoint-2880/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..33f73d5e9c8deebc108bef385e5e74b1ef8c8542
--- /dev/null
+++ b/runs/baseline/checkpoint-2880/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:721348485e694b0135ed2c93c3d0877d9feeaaee742aa29f66371f9eba4ec67c
+size 583356232
diff --git a/runs/baseline/checkpoint-2880/optimizer.pt b/runs/baseline/checkpoint-2880/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..9f30555c64a5f60c1d47dcc457c7bdce06ac3611
--- /dev/null
+++ b/runs/baseline/checkpoint-2880/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:820cb1774150899abf0c334ccf7b12827b985cc98b69314900a502c06630797e
+size 1166825803
diff --git a/runs/baseline/checkpoint-2880/rng_state_0.pth b/runs/baseline/checkpoint-2880/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..b73f9a380e4e9c61520ae0674c9453be2b007f8d
--- /dev/null
+++ b/runs/baseline/checkpoint-2880/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:19e0f78f2d9435829a3e127bfda2ca14f3245c5620079efefde656536bbe279e
+size 14917
diff --git a/runs/baseline/checkpoint-2880/rng_state_1.pth b/runs/baseline/checkpoint-2880/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..1c80ff79c801a37a1192dfdb25ad4404ef9b73aa
--- /dev/null
+++ b/runs/baseline/checkpoint-2880/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0ff567f23d2e2493cacddea05b5da8eee7e39fe5f45b9a2a98aa6a743f71addc
+size 14917
diff --git a/runs/baseline/checkpoint-2880/scheduler.pt b/runs/baseline/checkpoint-2880/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..fe8f25d9d984b4f8840e0a426b55887e7fc8a096
--- /dev/null
+++ b/runs/baseline/checkpoint-2880/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c98bbcd7f4a448da5d1b7f87739b30169f3b0d703d298db87b75054e6b96f54a
+size 1465
diff --git a/runs/baseline/checkpoint-2880/tokenizer.json b/runs/baseline/checkpoint-2880/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/baseline/checkpoint-2880/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/baseline/checkpoint-2880/tokenizer_config.json b/runs/baseline/checkpoint-2880/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/baseline/checkpoint-2880/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/baseline/checkpoint-2880/trainer_state.json b/runs/baseline/checkpoint-2880/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..461d29f690c3fe5229d1fba4311e1b133e9ad6dc
--- /dev/null
+++ b/runs/baseline/checkpoint-2880/trainer_state.json
@@ -0,0 +1,20506 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 40.0,
+ "eval_steps": 500,
+ "global_step": 2880,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.12165174633264542,
+ "learning_rate": 0.0,
+ "loss": 12.027063369750977,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.12109383195638657,
+ "learning_rate": 3e-05,
+ "loss": 12.028635025024414,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.12690064311027527,
+ "learning_rate": 6e-05,
+ "loss": 11.894524574279785,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.1537286937236786,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.672262191772461,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.11778244376182556,
+ "learning_rate": 0.00012,
+ "loss": 11.433671951293945,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.10874935984611511,
+ "learning_rate": 0.00015000000000000001,
+ "loss": 11.270309448242188,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.10641921311616898,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 11.155481338500977,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.1048506498336792,
+ "learning_rate": 0.00021000000000000004,
+ "loss": 11.066675186157227,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.10546834021806717,
+ "learning_rate": 0.00024,
+ "loss": 10.952295303344727,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.10614268481731415,
+ "learning_rate": 0.00027,
+ "loss": 10.811386108398438,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10529010742902756,
+ "learning_rate": 0.00030000000000000003,
+ "loss": 10.657712936401367,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.1051444336771965,
+ "learning_rate": 0.00033,
+ "loss": 10.47591781616211,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10420846194028854,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 10.280141830444336,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10323625057935715,
+ "learning_rate": 0.00039000000000000005,
+ "loss": 10.076366424560547,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10255401581525803,
+ "learning_rate": 0.00042000000000000007,
+ "loss": 9.856939315795898,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10151027888059616,
+ "learning_rate": 0.00045,
+ "loss": 9.632134437561035,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.1006326898932457,
+ "learning_rate": 0.00048,
+ "loss": 9.401637077331543,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.09820562601089478,
+ "learning_rate": 0.00051,
+ "loss": 9.177400588989258,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.0956626906991005,
+ "learning_rate": 0.00054,
+ "loss": 8.95518684387207,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.0922091156244278,
+ "learning_rate": 0.00057,
+ "loss": 8.75066089630127,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.08593932539224625,
+ "learning_rate": 0.0006000000000000001,
+ "loss": 8.557976722717285,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.08014766871929169,
+ "learning_rate": 0.00063,
+ "loss": 8.360604286193848,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.07381244748830795,
+ "learning_rate": 0.00066,
+ "loss": 8.210375785827637,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.06455052644014359,
+ "learning_rate": 0.0006900000000000001,
+ "loss": 8.07168960571289,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.054115474224090576,
+ "learning_rate": 0.0007199999999999999,
+ "loss": 7.960069179534912,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.043575264513492584,
+ "learning_rate": 0.00075,
+ "loss": 7.850685119628906,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.02976679615676403,
+ "learning_rate": 0.0007800000000000001,
+ "loss": 7.824377059936523,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.020552165806293488,
+ "learning_rate": 0.0008100000000000001,
+ "loss": 7.7695417404174805,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.018530743196606636,
+ "learning_rate": 0.0008400000000000001,
+ "loss": 7.782247543334961,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.02716820128262043,
+ "learning_rate": 0.00087,
+ "loss": 7.754945278167725,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.03638409823179245,
+ "learning_rate": 0.0009,
+ "loss": 7.799381256103516,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.041579652577638626,
+ "learning_rate": 0.00093,
+ "loss": 7.776629447937012,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.0304960235953331,
+ "learning_rate": 0.00096,
+ "loss": 7.749439239501953,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.053728822618722916,
+ "learning_rate": 0.00099,
+ "loss": 7.745059490203857,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.036322493106126785,
+ "learning_rate": 0.00102,
+ "loss": 7.752074718475342,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.03232959285378456,
+ "learning_rate": 0.00105,
+ "loss": 7.704257011413574,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.03246184438467026,
+ "learning_rate": 0.00108,
+ "loss": 7.667045593261719,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.036655545234680176,
+ "learning_rate": 0.00111,
+ "loss": 7.623445987701416,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.021042725071310997,
+ "learning_rate": 0.00114,
+ "loss": 7.582821846008301,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.022982975468039513,
+ "learning_rate": 0.00117,
+ "loss": 7.58036994934082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.024781864136457443,
+ "learning_rate": 0.0012000000000000001,
+ "loss": 7.5903000831604,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.03667089343070984,
+ "learning_rate": 0.00123,
+ "loss": 7.588233947753906,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.04083017632365227,
+ "learning_rate": 0.00126,
+ "loss": 7.548343658447266,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.02585385926067829,
+ "learning_rate": 0.00129,
+ "loss": 7.604833602905273,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.018793931230902672,
+ "learning_rate": 0.00132,
+ "loss": 7.56705379486084,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.025538936257362366,
+ "learning_rate": 0.00135,
+ "loss": 7.562107086181641,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.02819318138062954,
+ "learning_rate": 0.0013800000000000002,
+ "loss": 7.517213821411133,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.016239255666732788,
+ "learning_rate": 0.00141,
+ "loss": 7.4780755043029785,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.024470701813697815,
+ "learning_rate": 0.0014399999999999999,
+ "loss": 7.505887985229492,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.015499847941100597,
+ "learning_rate": 0.00147,
+ "loss": 7.484469890594482,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.017785200849175453,
+ "learning_rate": 0.0015,
+ "loss": 7.481822967529297,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.015952805057168007,
+ "learning_rate": 0.0015300000000000001,
+ "loss": 7.484804153442383,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.015771260485053062,
+ "learning_rate": 0.0015600000000000002,
+ "loss": 7.4495439529418945,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.024370530620217323,
+ "learning_rate": 0.00159,
+ "loss": 7.471208572387695,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.018173737451434135,
+ "learning_rate": 0.0016200000000000001,
+ "loss": 7.438447952270508,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.020564094185829163,
+ "learning_rate": 0.0016500000000000002,
+ "loss": 7.437922954559326,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.016154944896697998,
+ "learning_rate": 0.0016800000000000003,
+ "loss": 7.429378032684326,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.016817007213830948,
+ "learning_rate": 0.00171,
+ "loss": 7.406428337097168,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.03460947796702385,
+ "learning_rate": 0.00174,
+ "loss": 7.393331527709961,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.15410488843917847,
+ "learning_rate": 0.0017699999999999999,
+ "loss": 7.645856857299805,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.07443148642778397,
+ "learning_rate": 0.0018,
+ "loss": 7.549508094787598,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.0507572740316391,
+ "learning_rate": 0.00183,
+ "loss": 7.517921447753906,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.09682746231555939,
+ "learning_rate": 0.00186,
+ "loss": 7.543978691101074,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.2571517527103424,
+ "learning_rate": 0.00189,
+ "loss": 7.768949508666992,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.0317784883081913,
+ "learning_rate": 0.00192,
+ "loss": 7.509063720703125,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.0353970006108284,
+ "learning_rate": 0.0019500000000000001,
+ "loss": 7.522322654724121,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.03663994371891022,
+ "learning_rate": 0.00198,
+ "loss": 7.504723072052002,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.025481393560767174,
+ "learning_rate": 0.00201,
+ "loss": 7.452597141265869,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.03099915198981762,
+ "learning_rate": 0.00204,
+ "loss": 7.455441474914551,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.022921698167920113,
+ "learning_rate": 0.00207,
+ "loss": 7.4454827308654785,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.032388731837272644,
+ "learning_rate": 0.0021,
+ "loss": 7.467103958129883,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.022428609430789948,
+ "learning_rate": 0.00213,
+ "loss": 7.467540264129639,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.469719409942627,
+ "eval_runtime": 50.9714,
+ "eval_samples_per_second": 47.909,
+ "eval_steps_per_second": 1.511,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.017480911687016487,
+ "learning_rate": 0.00216,
+ "loss": 7.455810546875,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.016675131395459175,
+ "learning_rate": 0.00219,
+ "loss": 7.474848747253418,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.017450228333473206,
+ "learning_rate": 0.00222,
+ "loss": 7.442611217498779,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.016823675483465195,
+ "learning_rate": 0.0022500000000000003,
+ "loss": 7.454628944396973,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.016717543825507164,
+ "learning_rate": 0.00228,
+ "loss": 7.425317764282227,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.01644429750740528,
+ "learning_rate": 0.00231,
+ "loss": 7.4102864265441895,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.013104400597512722,
+ "learning_rate": 0.00234,
+ "loss": 7.4559173583984375,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.01175684668123722,
+ "learning_rate": 0.00237,
+ "loss": 7.438898086547852,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.013736684806644917,
+ "learning_rate": 0.0024000000000000002,
+ "loss": 7.449005603790283,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01176756713539362,
+ "learning_rate": 0.0024300000000000003,
+ "loss": 7.428421974182129,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.011860918253660202,
+ "learning_rate": 0.00246,
+ "loss": 7.411491394042969,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.010093477554619312,
+ "learning_rate": 0.00249,
+ "loss": 7.39311408996582,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.009453288279473782,
+ "learning_rate": 0.00252,
+ "loss": 7.387656211853027,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.010311692021787167,
+ "learning_rate": 0.00255,
+ "loss": 7.401076316833496,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.010053195990622044,
+ "learning_rate": 0.00258,
+ "loss": 7.39655876159668,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.009758037514984608,
+ "learning_rate": 0.00261,
+ "loss": 7.423122882843018,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008394946344196796,
+ "learning_rate": 0.00264,
+ "loss": 7.383601188659668,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.008457187563180923,
+ "learning_rate": 0.00267,
+ "loss": 7.400934219360352,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.010601334273815155,
+ "learning_rate": 0.0027,
+ "loss": 7.400241851806641,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.009009373374283314,
+ "learning_rate": 0.0027300000000000002,
+ "loss": 7.377348899841309,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.00737508200109005,
+ "learning_rate": 0.0027600000000000003,
+ "loss": 7.385881423950195,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.0075711836107075214,
+ "learning_rate": 0.0027900000000000004,
+ "loss": 7.348570823669434,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.00846901722252369,
+ "learning_rate": 0.00282,
+ "loss": 7.364287853240967,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.007160635199397802,
+ "learning_rate": 0.00285,
+ "loss": 7.3881635665893555,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.00682067358866334,
+ "learning_rate": 0.0028799999999999997,
+ "loss": 7.348663330078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.009108101949095726,
+ "learning_rate": 0.00291,
+ "loss": 7.356302261352539,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.015828927978873253,
+ "learning_rate": 0.00294,
+ "loss": 7.347576141357422,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.009683096781373024,
+ "learning_rate": 0.00297,
+ "loss": 7.323967933654785,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.010833270847797394,
+ "learning_rate": 0.003,
+ "loss": 7.3139495849609375,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.014905400574207306,
+ "learning_rate": 0.002999999401781161,
+ "loss": 7.302030086517334,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.023674853146076202,
+ "learning_rate": 0.002999997607125127,
+ "loss": 7.311811923980713,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.050460800528526306,
+ "learning_rate": 0.002999994616033342,
+ "loss": 7.41972541809082,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.02193174883723259,
+ "learning_rate": 0.002999990428508218,
+ "loss": 7.294468879699707,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.036261629313230515,
+ "learning_rate": 0.0029999850445531283,
+ "loss": 7.295568466186523,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.02734929881989956,
+ "learning_rate": 0.002999978464172409,
+ "loss": 7.309239864349365,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.017884107306599617,
+ "learning_rate": 0.002999970687371363,
+ "loss": 7.2427592277526855,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01700308918952942,
+ "learning_rate": 0.0029999617141562573,
+ "loss": 7.241730213165283,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.0209935512393713,
+ "learning_rate": 0.0029999515445343194,
+ "loss": 7.262085437774658,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.023207105696201324,
+ "learning_rate": 0.0029999401785137437,
+ "loss": 7.25450325012207,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.01907353848218918,
+ "learning_rate": 0.002999927616103687,
+ "loss": 7.23149299621582,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.016149666160345078,
+ "learning_rate": 0.0029999138573142715,
+ "loss": 7.210232734680176,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.017513928934931755,
+ "learning_rate": 0.002999898902156582,
+ "loss": 7.218451976776123,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.015002547763288021,
+ "learning_rate": 0.002999882750642667,
+ "loss": 7.169316291809082,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.014233081601560116,
+ "learning_rate": 0.0029998654027855406,
+ "loss": 7.194825172424316,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.01283260714262724,
+ "learning_rate": 0.0029998468585991783,
+ "loss": 7.15669059753418,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.014416799880564213,
+ "learning_rate": 0.0029998271180985223,
+ "loss": 7.143369674682617,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.014337058179080486,
+ "learning_rate": 0.0029998061812994757,
+ "loss": 7.127734184265137,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.016406210139393806,
+ "learning_rate": 0.002999784048218908,
+ "loss": 7.086504936218262,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.01680421642959118,
+ "learning_rate": 0.002999760718874652,
+ "loss": 7.089420318603516,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.010665548034012318,
+ "learning_rate": 0.0029997361932855016,
+ "loss": 7.102463722229004,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013479137793183327,
+ "learning_rate": 0.002999710471471218,
+ "loss": 7.079098701477051,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.012408486567437649,
+ "learning_rate": 0.0029996835534525246,
+ "loss": 7.0790114402771,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.012035196647047997,
+ "learning_rate": 0.0029996554392511094,
+ "loss": 7.058756351470947,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.00899985246360302,
+ "learning_rate": 0.0029996261288896225,
+ "loss": 7.0739240646362305,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.01072906143963337,
+ "learning_rate": 0.002999595622391679,
+ "loss": 7.085537910461426,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.01047342736274004,
+ "learning_rate": 0.0029995639197818577,
+ "loss": 7.060049057006836,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.02342618815600872,
+ "learning_rate": 0.0029995310210857006,
+ "loss": 7.039482116699219,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.011256859637796879,
+ "learning_rate": 0.0029994969263297134,
+ "loss": 7.040412902832031,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.008407019078731537,
+ "learning_rate": 0.0029994616355413666,
+ "loss": 7.0364274978637695,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.011006148532032967,
+ "learning_rate": 0.0029994251487490916,
+ "loss": 6.991468906402588,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.009103255346417427,
+ "learning_rate": 0.0029993874659822875,
+ "loss": 7.001059532165527,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.007875788025557995,
+ "learning_rate": 0.0029993485872713125,
+ "loss": 6.9974260330200195,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.007872611284255981,
+ "learning_rate": 0.0029993085126474913,
+ "loss": 6.980216979980469,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.010071957483887672,
+ "learning_rate": 0.0029992672421431115,
+ "loss": 6.97799015045166,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.01209618803113699,
+ "learning_rate": 0.0029992247757914237,
+ "loss": 6.97202205657959,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.014884858392179012,
+ "learning_rate": 0.0029991811136266425,
+ "loss": 6.976570129394531,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.014953463338315487,
+ "learning_rate": 0.0029991362556839453,
+ "loss": 6.971050262451172,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.008371063508093357,
+ "learning_rate": 0.002999090201999474,
+ "loss": 6.955556869506836,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.011926881037652493,
+ "learning_rate": 0.002999042952610332,
+ "loss": 6.943947792053223,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.013422893360257149,
+ "learning_rate": 0.002998994507554589,
+ "loss": 6.910022735595703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.009672224521636963,
+ "learning_rate": 0.002998944866871275,
+ "loss": 6.8994855880737305,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.012590581551194191,
+ "learning_rate": 0.0029988940306003853,
+ "loss": 6.88709020614624,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.945146083831787,
+ "eval_runtime": 51.5801,
+ "eval_samples_per_second": 47.344,
+ "eval_steps_per_second": 1.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.016554836183786392,
+ "learning_rate": 0.002998841998782877,
+ "loss": 6.916609764099121,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.011085660196840763,
+ "learning_rate": 0.0029987887714606723,
+ "loss": 6.884910583496094,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.008596168830990791,
+ "learning_rate": 0.0029987343486766547,
+ "loss": 6.900066375732422,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.014487302862107754,
+ "learning_rate": 0.0029986787304746715,
+ "loss": 6.8693342208862305,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.015636997297406197,
+ "learning_rate": 0.002998621916899534,
+ "loss": 6.903670787811279,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.013275462202727795,
+ "learning_rate": 0.0029985639079970153,
+ "loss": 6.889361381530762,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.009044744074344635,
+ "learning_rate": 0.002998504703813852,
+ "loss": 6.8549981117248535,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.011327696964144707,
+ "learning_rate": 0.002998444304397745,
+ "loss": 6.856151103973389,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.01310331467539072,
+ "learning_rate": 0.0029983827097973554,
+ "loss": 6.834944725036621,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.010003753937780857,
+ "learning_rate": 0.00299831992006231,
+ "loss": 6.825613975524902,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.009928115643560886,
+ "learning_rate": 0.0029982559352431967,
+ "loss": 6.828935146331787,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.012302150949835777,
+ "learning_rate": 0.0029981907553915674,
+ "loss": 6.821207046508789,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.020555168390274048,
+ "learning_rate": 0.002998124380559936,
+ "loss": 6.8184895515441895,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.011417638510465622,
+ "learning_rate": 0.0029980568108017797,
+ "loss": 6.788754463195801,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.010772978886961937,
+ "learning_rate": 0.002997988046171538,
+ "loss": 6.7861433029174805,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.012753859162330627,
+ "learning_rate": 0.002997918086724614,
+ "loss": 6.788776397705078,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.011327114887535572,
+ "learning_rate": 0.0029978469325173717,
+ "loss": 6.779160022735596,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01235753484070301,
+ "learning_rate": 0.002997774583607139,
+ "loss": 6.799554824829102,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.01417156495153904,
+ "learning_rate": 0.0029977010400522074,
+ "loss": 6.77957820892334,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.021321553736925125,
+ "learning_rate": 0.002997626301911828,
+ "loss": 6.785632133483887,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.03614472970366478,
+ "learning_rate": 0.002997550369246217,
+ "loss": 6.836414337158203,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03934052586555481,
+ "learning_rate": 0.002997473242116551,
+ "loss": 6.854659557342529,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.016240060329437256,
+ "learning_rate": 0.0029973949205849707,
+ "loss": 6.789690971374512,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.031164905056357384,
+ "learning_rate": 0.0029973154047145782,
+ "loss": 6.8363776206970215,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.013051844201982021,
+ "learning_rate": 0.002997234694569438,
+ "loss": 6.803576469421387,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.022611448541283607,
+ "learning_rate": 0.002997152790214577,
+ "loss": 6.762107849121094,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.017812907695770264,
+ "learning_rate": 0.0029970696917159833,
+ "loss": 6.776218414306641,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.015453332103788853,
+ "learning_rate": 0.0029969853991406088,
+ "loss": 6.742618083953857,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.020630091428756714,
+ "learning_rate": 0.0029968999125563655,
+ "loss": 6.765475749969482,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.014672341756522655,
+ "learning_rate": 0.0029968132320321297,
+ "loss": 6.754009246826172,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.016540220007300377,
+ "learning_rate": 0.0029967253576377376,
+ "loss": 6.734752655029297,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.013495649211108685,
+ "learning_rate": 0.0029966362894439874,
+ "loss": 6.752676010131836,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015936989337205887,
+ "learning_rate": 0.0029965460275226405,
+ "loss": 6.725708961486816,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010431243106722832,
+ "learning_rate": 0.0029964545719464194,
+ "loss": 6.698376655578613,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01669321395456791,
+ "learning_rate": 0.0029963619227890083,
+ "loss": 6.711770534515381,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.013308997265994549,
+ "learning_rate": 0.0029962680801250514,
+ "loss": 6.705286979675293,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011648476123809814,
+ "learning_rate": 0.002996173044030159,
+ "loss": 6.690052509307861,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.03942161798477173,
+ "learning_rate": 0.0029960768145808965,
+ "loss": 6.711885452270508,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.01571274921298027,
+ "learning_rate": 0.002995979391854797,
+ "loss": 6.715847969055176,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.00829714722931385,
+ "learning_rate": 0.0029958807759303515,
+ "loss": 6.685308933258057,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.01374727487564087,
+ "learning_rate": 0.0029957809668870114,
+ "loss": 6.695744514465332,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.010857059620320797,
+ "learning_rate": 0.0029956799648051934,
+ "loss": 6.658294677734375,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.008897246792912483,
+ "learning_rate": 0.002995577769766272,
+ "loss": 6.642439842224121,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015880772843956947,
+ "learning_rate": 0.002995474381852584,
+ "loss": 6.674907684326172,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.021623898297548294,
+ "learning_rate": 0.0029953698011474265,
+ "loss": 6.66236686706543,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.03084627166390419,
+ "learning_rate": 0.0029952640277350596,
+ "loss": 6.666715621948242,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.040561504662036896,
+ "learning_rate": 0.002995157061700702,
+ "loss": 6.649062156677246,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.031126515939831734,
+ "learning_rate": 0.0029950489031305345,
+ "loss": 6.640542030334473,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.00872360821813345,
+ "learning_rate": 0.0029949395521116987,
+ "loss": 6.635221481323242,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.023432321846485138,
+ "learning_rate": 0.0029948290087322966,
+ "loss": 6.642196178436279,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.021096013486385345,
+ "learning_rate": 0.0029947172730813913,
+ "loss": 6.643655300140381,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.014494858682155609,
+ "learning_rate": 0.002994604345249006,
+ "loss": 6.612720966339111,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.05225599184632301,
+ "learning_rate": 0.0029944902253261248,
+ "loss": 6.60659646987915,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.014311269856989384,
+ "learning_rate": 0.0029943749134046915,
+ "loss": 6.54373025894165,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.010777472518384457,
+ "learning_rate": 0.0029942584095776115,
+ "loss": 6.599837303161621,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.012498079799115658,
+ "learning_rate": 0.002994140713938749,
+ "loss": 6.589657783508301,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.015775132924318314,
+ "learning_rate": 0.00299402182658293,
+ "loss": 6.577511787414551,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.017978230491280556,
+ "learning_rate": 0.0029939017476059396,
+ "loss": 6.58717155456543,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.026807745918631554,
+ "learning_rate": 0.0029937804771045236,
+ "loss": 6.578145503997803,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.04853282868862152,
+ "learning_rate": 0.0029936580151763872,
+ "loss": 6.609755516052246,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.1057916209101677,
+ "learning_rate": 0.002993534361920195,
+ "loss": 6.870367050170898,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.04925911873579025,
+ "learning_rate": 0.002993409517435574,
+ "loss": 6.871556282043457,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.05288556218147278,
+ "learning_rate": 0.0029932834818231077,
+ "loss": 6.808315753936768,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.03831710293889046,
+ "learning_rate": 0.0029931562551843405,
+ "loss": 6.827617645263672,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.03665117174386978,
+ "learning_rate": 0.002993027837621778,
+ "loss": 6.767284870147705,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.026216736063361168,
+ "learning_rate": 0.0029928982292388826,
+ "loss": 6.733659744262695,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.024796854704618454,
+ "learning_rate": 0.002992767430140078,
+ "loss": 6.759577751159668,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.02410992607474327,
+ "learning_rate": 0.002992635440430747,
+ "loss": 6.701825141906738,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.019924134016036987,
+ "learning_rate": 0.0029925022602172308,
+ "loss": 6.677121162414551,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.022516867145895958,
+ "learning_rate": 0.00299236788960683,
+ "loss": 6.668622970581055,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.025068556889891624,
+ "learning_rate": 0.002992232328707806,
+ "loss": 6.67718505859375,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.02574208751320839,
+ "learning_rate": 0.002992095577629376,
+ "loss": 6.679594039916992,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.657555103302002,
+ "eval_runtime": 51.0199,
+ "eval_samples_per_second": 47.864,
+ "eval_steps_per_second": 1.509,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.017298126593232155,
+ "learning_rate": 0.0029919576364817184,
+ "loss": 6.620774269104004,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.018840325996279716,
+ "learning_rate": 0.002991818505375971,
+ "loss": 6.613316535949707,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.020813113078475,
+ "learning_rate": 0.002991678184424228,
+ "loss": 6.621555328369141,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.026581479236483574,
+ "learning_rate": 0.0029915366737395437,
+ "loss": 6.613471031188965,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.02084030583500862,
+ "learning_rate": 0.0029913939734359306,
+ "loss": 6.559704780578613,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.01327502354979515,
+ "learning_rate": 0.00299125008362836,
+ "loss": 6.564910888671875,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.015091734007000923,
+ "learning_rate": 0.0029911050044327614,
+ "loss": 6.571262836456299,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.015463982708752155,
+ "learning_rate": 0.002990958735966023,
+ "loss": 6.565750598907471,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.013636826537549496,
+ "learning_rate": 0.0029908112783459894,
+ "loss": 6.57981014251709,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.010613273829221725,
+ "learning_rate": 0.0029906626316914655,
+ "loss": 6.538505554199219,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.01804237812757492,
+ "learning_rate": 0.0029905127961222127,
+ "loss": 6.5282511711120605,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.015131509862840176,
+ "learning_rate": 0.0029903617717589506,
+ "loss": 6.531335830688477,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.010518206283450127,
+ "learning_rate": 0.002990209558723358,
+ "loss": 6.521886825561523,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.011360022239387035,
+ "learning_rate": 0.002990056157138069,
+ "loss": 6.528301239013672,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.011556181125342846,
+ "learning_rate": 0.0029899015671266773,
+ "loss": 6.497917652130127,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.01365791168063879,
+ "learning_rate": 0.0029897457888137325,
+ "loss": 6.52849006652832,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.01091043371707201,
+ "learning_rate": 0.0029895888223247426,
+ "loss": 6.492197036743164,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.008484847843647003,
+ "learning_rate": 0.002989430667786173,
+ "loss": 6.497377395629883,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.00836879201233387,
+ "learning_rate": 0.0029892713253254466,
+ "loss": 6.495189666748047,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.009781497530639172,
+ "learning_rate": 0.0029891107950709406,
+ "loss": 6.454085826873779,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.009603648446500301,
+ "learning_rate": 0.0029889490771519934,
+ "loss": 6.499475479125977,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.007991855964064598,
+ "learning_rate": 0.0029887861716988975,
+ "loss": 6.460785865783691,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.008447187952697277,
+ "learning_rate": 0.0029886220788429025,
+ "loss": 6.504582405090332,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.010945498012006283,
+ "learning_rate": 0.0029884567987162156,
+ "loss": 6.49589729309082,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.011096215806901455,
+ "learning_rate": 0.0029882903314519997,
+ "loss": 6.460657119750977,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.013094739057123661,
+ "learning_rate": 0.0029881226771843747,
+ "loss": 6.462450981140137,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.015423037111759186,
+ "learning_rate": 0.0029879538360484165,
+ "loss": 6.45565128326416,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.0182222630828619,
+ "learning_rate": 0.002987783808180157,
+ "loss": 6.452227592468262,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.01981227844953537,
+ "learning_rate": 0.002987612593716585,
+ "loss": 6.477701663970947,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.026733700186014175,
+ "learning_rate": 0.002987440192795645,
+ "loss": 6.452212333679199,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.03504155948758125,
+ "learning_rate": 0.0029872666055562372,
+ "loss": 6.509665489196777,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.027970125898718834,
+ "learning_rate": 0.002987091832138217,
+ "loss": 6.487089157104492,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.013714591972529888,
+ "learning_rate": 0.0029869158726823974,
+ "loss": 6.469569206237793,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.016472695395350456,
+ "learning_rate": 0.0029867387273305453,
+ "loss": 6.461688995361328,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.018457358703017235,
+ "learning_rate": 0.0029865603962253833,
+ "loss": 6.490985870361328,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.011045100167393684,
+ "learning_rate": 0.0029863808795105896,
+ "loss": 6.403701305389404,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.014126897789537907,
+ "learning_rate": 0.0029862001773307972,
+ "loss": 6.379804611206055,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.014015335589647293,
+ "learning_rate": 0.0029860182898315954,
+ "loss": 6.404377460479736,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.010032658465206623,
+ "learning_rate": 0.002985835217159527,
+ "loss": 6.40291690826416,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.015469838865101337,
+ "learning_rate": 0.0029856509594620912,
+ "loss": 6.4414520263671875,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.011106910184025764,
+ "learning_rate": 0.0029854655168877397,
+ "loss": 6.396444320678711,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.01262320950627327,
+ "learning_rate": 0.0029852788895858816,
+ "loss": 6.385222434997559,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.012783976271748543,
+ "learning_rate": 0.002985091077706878,
+ "loss": 6.3818135261535645,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.01224067248404026,
+ "learning_rate": 0.0029849020814020468,
+ "loss": 6.37886905670166,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.010540086776018143,
+ "learning_rate": 0.0029847119008236576,
+ "loss": 6.37799072265625,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.012479913420975208,
+ "learning_rate": 0.0029845205361249366,
+ "loss": 6.365094184875488,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.010080077685415745,
+ "learning_rate": 0.0029843279874600622,
+ "loss": 6.371852874755859,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.013266298919916153,
+ "learning_rate": 0.002984134254984168,
+ "loss": 6.380039215087891,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.01014771405607462,
+ "learning_rate": 0.002983939338853339,
+ "loss": 6.370337963104248,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.008164579048752785,
+ "learning_rate": 0.0029837432392246185,
+ "loss": 6.331899642944336,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.01020442508161068,
+ "learning_rate": 0.002983545956255998,
+ "loss": 6.35077428817749,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.009898579679429531,
+ "learning_rate": 0.002983347490106426,
+ "loss": 6.371985912322998,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.009668919257819653,
+ "learning_rate": 0.002983147840935802,
+ "loss": 6.356776237487793,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.012560366652905941,
+ "learning_rate": 0.0029829470089049815,
+ "loss": 6.357911109924316,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.018435249105095863,
+ "learning_rate": 0.0029827449941757698,
+ "loss": 6.366821765899658,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.029046855866909027,
+ "learning_rate": 0.002982541796910927,
+ "loss": 6.367595672607422,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.043010637164115906,
+ "learning_rate": 0.002982337417274165,
+ "loss": 6.413588523864746,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.045470818877220154,
+ "learning_rate": 0.0029821318554301497,
+ "loss": 6.435015678405762,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.02148726023733616,
+ "learning_rate": 0.002981925111544498,
+ "loss": 6.330894470214844,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.026808688417077065,
+ "learning_rate": 0.00298171718578378,
+ "loss": 6.384839057922363,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.018495213240385056,
+ "learning_rate": 0.002981508078315517,
+ "loss": 6.36277961730957,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.025393597781658173,
+ "learning_rate": 0.0029812977893081834,
+ "loss": 6.3702239990234375,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020856289193034172,
+ "learning_rate": 0.002981086318931206,
+ "loss": 6.354547500610352,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.017827816307544708,
+ "learning_rate": 0.002980873667354962,
+ "loss": 6.339264869689941,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.02639714442193508,
+ "learning_rate": 0.002980659834750781,
+ "loss": 6.3254194259643555,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.015238096937537193,
+ "learning_rate": 0.002980444821290944,
+ "loss": 6.365842819213867,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.012863729149103165,
+ "learning_rate": 0.002980228627148684,
+ "loss": 6.356776714324951,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.014739899896085262,
+ "learning_rate": 0.0029800112524981845,
+ "loss": 6.332481384277344,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.013582171872258186,
+ "learning_rate": 0.0029797926975145797,
+ "loss": 6.311968803405762,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.012704637832939625,
+ "learning_rate": 0.0029795729623739555,
+ "loss": 6.321654319763184,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.012550560757517815,
+ "learning_rate": 0.0029793520472533495,
+ "loss": 6.298409461975098,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.011170165613293648,
+ "learning_rate": 0.002979129952330748,
+ "loss": 6.294830799102783,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.326757907867432,
+ "eval_runtime": 51.6566,
+ "eval_samples_per_second": 47.274,
+ "eval_steps_per_second": 1.491,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.010947446338832378,
+ "learning_rate": 0.0029789066777850897,
+ "loss": 6.286581039428711,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.010097995400428772,
+ "learning_rate": 0.002978682223796262,
+ "loss": 6.288301467895508,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.01224475260823965,
+ "learning_rate": 0.002978456590545104,
+ "loss": 6.264986991882324,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.009734537452459335,
+ "learning_rate": 0.0029782297782134045,
+ "loss": 6.304243087768555,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.01136956550180912,
+ "learning_rate": 0.0029780017869839013,
+ "loss": 6.244938373565674,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.00935293547809124,
+ "learning_rate": 0.002977772617040284,
+ "loss": 6.251852035522461,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.011118784546852112,
+ "learning_rate": 0.0029775422685671894,
+ "loss": 6.256633758544922,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.010837923735380173,
+ "learning_rate": 0.0029773107417502055,
+ "loss": 6.263411045074463,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.011920305900275707,
+ "learning_rate": 0.0029770780367758703,
+ "loss": 6.239311218261719,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.014117483049631119,
+ "learning_rate": 0.0029768441538316685,
+ "loss": 6.253314018249512,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.022502148523926735,
+ "learning_rate": 0.002976609093106036,
+ "loss": 6.246156692504883,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.03348720073699951,
+ "learning_rate": 0.002976372854788357,
+ "loss": 6.255914688110352,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.03427121415734291,
+ "learning_rate": 0.002976135439068965,
+ "loss": 6.289946556091309,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.02407355234026909,
+ "learning_rate": 0.002975896846139141,
+ "loss": 6.288534164428711,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.017861567437648773,
+ "learning_rate": 0.0029756570761911147,
+ "loss": 6.258049964904785,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.019980227574706078,
+ "learning_rate": 0.0029754161294180646,
+ "loss": 6.250895023345947,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.01443127728998661,
+ "learning_rate": 0.0029751740060141177,
+ "loss": 6.212200164794922,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.014975055120885372,
+ "learning_rate": 0.0029749307061743475,
+ "loss": 6.230568885803223,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01408296450972557,
+ "learning_rate": 0.0029746862300947773,
+ "loss": 6.243950366973877,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.011649327352643013,
+ "learning_rate": 0.0029744405779723766,
+ "loss": 6.197744846343994,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.013392188586294651,
+ "learning_rate": 0.002974193750005063,
+ "loss": 6.207277297973633,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.010902997106313705,
+ "learning_rate": 0.002973945746391701,
+ "loss": 6.166373252868652,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.012915392406284809,
+ "learning_rate": 0.0029736965673321032,
+ "loss": 6.186316013336182,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.010772830806672573,
+ "learning_rate": 0.0029734462130270273,
+ "loss": 6.172578811645508,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.010274911299347878,
+ "learning_rate": 0.002973194683678181,
+ "loss": 6.206851005554199,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.013210494071245193,
+ "learning_rate": 0.0029729419794882164,
+ "loss": 6.21115779876709,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.013372313231229782,
+ "learning_rate": 0.0029726881006607326,
+ "loss": 6.18442440032959,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.008105908520519733,
+ "learning_rate": 0.0029724330474002743,
+ "loss": 6.204354286193848,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.014167108573019505,
+ "learning_rate": 0.0029721768199123346,
+ "loss": 6.180662155151367,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.008339803665876389,
+ "learning_rate": 0.002971919418403351,
+ "loss": 6.150079727172852,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.015939386561512947,
+ "learning_rate": 0.0029716608430807073,
+ "loss": 6.175203323364258,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.010777109302580357,
+ "learning_rate": 0.0029714010941527327,
+ "loss": 6.166106224060059,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.011274735443294048,
+ "learning_rate": 0.002971140171828702,
+ "loss": 6.163111686706543,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013626808300614357,
+ "learning_rate": 0.002970878076318836,
+ "loss": 6.140850067138672,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.016907762736082077,
+ "learning_rate": 0.0029706148078343,
+ "loss": 6.192012786865234,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.018817441537976265,
+ "learning_rate": 0.002970350366587205,
+ "loss": 6.162238121032715,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.02829768694937229,
+ "learning_rate": 0.0029700847527906068,
+ "loss": 6.159066200256348,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.02034052088856697,
+ "learning_rate": 0.002969817966658505,
+ "loss": 6.164158344268799,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.01816369965672493,
+ "learning_rate": 0.0029695500084058447,
+ "loss": 6.132316589355469,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.013685261830687523,
+ "learning_rate": 0.0029692808782485146,
+ "loss": 6.152181625366211,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.009791149757802486,
+ "learning_rate": 0.0029690105764033485,
+ "loss": 6.138544082641602,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.014827069826424122,
+ "learning_rate": 0.002968739103088124,
+ "loss": 6.135766983032227,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.01391249243170023,
+ "learning_rate": 0.0029684664585215618,
+ "loss": 6.146172523498535,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.012729899026453495,
+ "learning_rate": 0.002968192642923326,
+ "loss": 6.122613430023193,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.014175458811223507,
+ "learning_rate": 0.002967917656514026,
+ "loss": 6.12574577331543,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.010958840139210224,
+ "learning_rate": 0.002967641499515213,
+ "loss": 6.111088752746582,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.015380816534161568,
+ "learning_rate": 0.0029673641721493817,
+ "loss": 6.100615501403809,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.025339096784591675,
+ "learning_rate": 0.00296708567463997,
+ "loss": 6.139659881591797,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.03695995733141899,
+ "learning_rate": 0.002966806007211358,
+ "loss": 6.160996437072754,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.0408046580851078,
+ "learning_rate": 0.0029665251700888686,
+ "loss": 6.173980712890625,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.025022024288773537,
+ "learning_rate": 0.0029662431634987674,
+ "loss": 6.145794868469238,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.027254154905676842,
+ "learning_rate": 0.002965959987668262,
+ "loss": 6.16928243637085,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.06191903352737427,
+ "learning_rate": 0.0029656756428255017,
+ "loss": 6.138525009155273,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0348956398665905,
+ "learning_rate": 0.002965390129199579,
+ "loss": 6.185477256774902,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.020859677344560623,
+ "learning_rate": 0.002965103447020526,
+ "loss": 6.148846626281738,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.022810814902186394,
+ "learning_rate": 0.0029648155965193174,
+ "loss": 6.1701178550720215,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.017981817945837975,
+ "learning_rate": 0.00296452657792787,
+ "loss": 6.1323771476745605,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.014905866235494614,
+ "learning_rate": 0.00296423639147904,
+ "loss": 6.123426914215088,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.01463468000292778,
+ "learning_rate": 0.002963945037406626,
+ "loss": 6.146122932434082,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.021616583690047264,
+ "learning_rate": 0.002963652515945366,
+ "loss": 6.093459606170654,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.01550968922674656,
+ "learning_rate": 0.00296335882733094,
+ "loss": 6.114929676055908,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.014597428031265736,
+ "learning_rate": 0.0029630639717999675,
+ "loss": 6.113945960998535,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.018249280750751495,
+ "learning_rate": 0.002962767949590008,
+ "loss": 6.130507469177246,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.014160148799419403,
+ "learning_rate": 0.0029624707609395614,
+ "loss": 6.086946487426758,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.015563832595944405,
+ "learning_rate": 0.002962172406088067,
+ "loss": 6.083332538604736,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.012240015901625156,
+ "learning_rate": 0.002961872885275904,
+ "loss": 6.0822906494140625,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.0141651201993227,
+ "learning_rate": 0.002961572198744391,
+ "loss": 6.049562454223633,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.010004645213484764,
+ "learning_rate": 0.0029612703467357866,
+ "loss": 6.067063808441162,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.011790835298597813,
+ "learning_rate": 0.0029609673294932856,
+ "loss": 6.062193870544434,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.011088060215115547,
+ "learning_rate": 0.002960663147261025,
+ "loss": 6.070950031280518,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.010387620888650417,
+ "learning_rate": 0.002960357800284078,
+ "loss": 6.032423973083496,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.012074458412826061,
+ "learning_rate": 0.002960051288808458,
+ "loss": 6.066788673400879,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.097111701965332,
+ "eval_runtime": 51.8742,
+ "eval_samples_per_second": 47.075,
+ "eval_steps_per_second": 1.484,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.013429952785372734,
+ "learning_rate": 0.0029597436130811153,
+ "loss": 6.022172451019287,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.01764705777168274,
+ "learning_rate": 0.0029594347733499387,
+ "loss": 6.029745101928711,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.02403099089860916,
+ "learning_rate": 0.002959124769863755,
+ "loss": 6.04707145690918,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.029030559584498405,
+ "learning_rate": 0.002958813602872328,
+ "loss": 6.05479097366333,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.029980644583702087,
+ "learning_rate": 0.0029585012726263595,
+ "loss": 6.068657875061035,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.021424276754260063,
+ "learning_rate": 0.0029581877793774885,
+ "loss": 6.062008857727051,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.018076395615935326,
+ "learning_rate": 0.002957873123378291,
+ "loss": 6.023279190063477,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.017688792198896408,
+ "learning_rate": 0.002957557304882279,
+ "loss": 6.014641284942627,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.02313626930117607,
+ "learning_rate": 0.0029572403241439026,
+ "loss": 6.026420593261719,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.04045062139630318,
+ "learning_rate": 0.0029569221814185467,
+ "loss": 6.007199764251709,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.015792209655046463,
+ "learning_rate": 0.0029566028769625334,
+ "loss": 6.017209053039551,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.01990479789674282,
+ "learning_rate": 0.002956282411033121,
+ "loss": 5.982841491699219,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.01191043108701706,
+ "learning_rate": 0.002955960783888503,
+ "loss": 6.0035929679870605,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.01732831820845604,
+ "learning_rate": 0.002955637995787807,
+ "loss": 6.025006294250488,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.032948266714811325,
+ "learning_rate": 0.0029553140469910995,
+ "loss": 6.007736682891846,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.011192727833986282,
+ "learning_rate": 0.00295498893775938,
+ "loss": 6.0022687911987305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014646840281784534,
+ "learning_rate": 0.0029546626683545824,
+ "loss": 5.976479530334473,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01589796133339405,
+ "learning_rate": 0.002954335239039576,
+ "loss": 5.980949878692627,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.015129734762012959,
+ "learning_rate": 0.0029540066500781657,
+ "loss": 6.0073394775390625,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.011140518821775913,
+ "learning_rate": 0.002953676901735088,
+ "loss": 5.976587295532227,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.0109480582177639,
+ "learning_rate": 0.0029533459942760174,
+ "loss": 6.001427173614502,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.013346790336072445,
+ "learning_rate": 0.0029530139279675583,
+ "loss": 5.96919059753418,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.012474581599235535,
+ "learning_rate": 0.002952680703077251,
+ "loss": 5.970426559448242,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01140566635876894,
+ "learning_rate": 0.002952346319873569,
+ "loss": 5.989171504974365,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.0137029392644763,
+ "learning_rate": 0.0029520107786259187,
+ "loss": 6.010115623474121,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015517761930823326,
+ "learning_rate": 0.002951674079604639,
+ "loss": 6.003649711608887,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.02173629403114319,
+ "learning_rate": 0.0029513362230810035,
+ "loss": 5.967074871063232,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02683914266526699,
+ "learning_rate": 0.002950997209327216,
+ "loss": 5.994434833526611,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.025865858420729637,
+ "learning_rate": 0.002950657038616415,
+ "loss": 6.010523796081543,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02544897235929966,
+ "learning_rate": 0.002950315711222668,
+ "loss": 5.994078159332275,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.028426019474864006,
+ "learning_rate": 0.0029499732274209777,
+ "loss": 6.020925521850586,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.026174750179052353,
+ "learning_rate": 0.0029496295874872768,
+ "loss": 5.985616683959961,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.02802138403058052,
+ "learning_rate": 0.00294928479169843,
+ "loss": 5.973659038543701,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.019348938018083572,
+ "learning_rate": 0.002948938840332233,
+ "loss": 5.990137577056885,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.016294041648507118,
+ "learning_rate": 0.0029485917336674115,
+ "loss": 5.962082862854004,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.016167785972356796,
+ "learning_rate": 0.002948243471983625,
+ "loss": 5.9594573974609375,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.014119232073426247,
+ "learning_rate": 0.00294789405556146,
+ "loss": 5.894467353820801,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.013805347494781017,
+ "learning_rate": 0.002947543484682436,
+ "loss": 5.964014530181885,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.014496497809886932,
+ "learning_rate": 0.002947191759629001,
+ "loss": 5.989920616149902,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.014813800342381,
+ "learning_rate": 0.0029468388806845333,
+ "loss": 5.9669413566589355,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.01153622381389141,
+ "learning_rate": 0.0029464848481333423,
+ "loss": 5.918018341064453,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.010525977239012718,
+ "learning_rate": 0.0029461296622606647,
+ "loss": 5.892477989196777,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.011410463601350784,
+ "learning_rate": 0.002945773323352667,
+ "loss": 5.925249099731445,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.010949424467980862,
+ "learning_rate": 0.0029454158316964464,
+ "loss": 5.90645694732666,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.00962804164737463,
+ "learning_rate": 0.002945057187580026,
+ "loss": 5.917054176330566,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.011517546139657497,
+ "learning_rate": 0.0029446973912923597,
+ "loss": 5.919445514678955,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01567666605114937,
+ "learning_rate": 0.0029443364431233285,
+ "loss": 5.884111404418945,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.016781140118837357,
+ "learning_rate": 0.002943974343363742,
+ "loss": 5.9252119064331055,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.02244546078145504,
+ "learning_rate": 0.0029436110923053375,
+ "loss": 5.933143615722656,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.028186792507767677,
+ "learning_rate": 0.00294324669024078,
+ "loss": 5.9079132080078125,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.028585223481059074,
+ "learning_rate": 0.002942881137463661,
+ "loss": 5.925262451171875,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.02781485952436924,
+ "learning_rate": 0.0029425144342685003,
+ "loss": 5.963315010070801,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.021486928686499596,
+ "learning_rate": 0.002942146580950744,
+ "loss": 5.930115699768066,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.016890818253159523,
+ "learning_rate": 0.002941777577806765,
+ "loss": 5.921439170837402,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.0169126745313406,
+ "learning_rate": 0.0029414074251338618,
+ "loss": 5.928220272064209,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.014659176580607891,
+ "learning_rate": 0.002941036123230261,
+ "loss": 5.934546947479248,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.015337883494794369,
+ "learning_rate": 0.002940663672395112,
+ "loss": 5.890145301818848,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01254988368600607,
+ "learning_rate": 0.002940290072928493,
+ "loss": 5.875492095947266,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.014658362604677677,
+ "learning_rate": 0.0029399153251314066,
+ "loss": 5.901272773742676,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.017342250794172287,
+ "learning_rate": 0.002939539429305779,
+ "loss": 5.875853538513184,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.016825608909130096,
+ "learning_rate": 0.0029391623857544643,
+ "loss": 5.941972732543945,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.016716884449124336,
+ "learning_rate": 0.002938784194781238,
+ "loss": 5.912367343902588,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.021707067266106606,
+ "learning_rate": 0.002938404856690803,
+ "loss": 5.895319938659668,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.030519135296344757,
+ "learning_rate": 0.002938024371788785,
+ "loss": 5.897165298461914,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.02572188340127468,
+ "learning_rate": 0.002937642740381733,
+ "loss": 5.879164695739746,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.016286717727780342,
+ "learning_rate": 0.002937259962777122,
+ "loss": 5.841421127319336,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.020656082779169083,
+ "learning_rate": 0.002936876039283347,
+ "loss": 5.9222259521484375,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.015350459143519402,
+ "learning_rate": 0.0029364909702097298,
+ "loss": 5.885847091674805,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01708570495247841,
+ "learning_rate": 0.0029361047558665136,
+ "loss": 5.872044563293457,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.013537376187741756,
+ "learning_rate": 0.0029357173965648634,
+ "loss": 5.866434097290039,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.013208637945353985,
+ "learning_rate": 0.002935328892616868,
+ "loss": 5.8624491691589355,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.012249554507434368,
+ "learning_rate": 0.002934939244335538,
+ "loss": 5.833232879638672,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.90985107421875,
+ "eval_runtime": 51.0696,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.01382362935692072,
+ "learning_rate": 0.0029345484520348062,
+ "loss": 5.835695266723633,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.012947693467140198,
+ "learning_rate": 0.0029341565160295262,
+ "loss": 5.826190948486328,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.013653872534632683,
+ "learning_rate": 0.002933763436635475,
+ "loss": 5.831789016723633,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.027596980333328247,
+ "learning_rate": 0.0029333692141693484,
+ "loss": 5.839963436126709,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01637990213930607,
+ "learning_rate": 0.0029329738489487648,
+ "loss": 5.805248260498047,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.02169182151556015,
+ "learning_rate": 0.0029325773412922627,
+ "loss": 5.838248252868652,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.03518201410770416,
+ "learning_rate": 0.0029321796915193015,
+ "loss": 5.834336280822754,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04983877018094063,
+ "learning_rate": 0.0029317808999502603,
+ "loss": 5.879856109619141,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05085664615035057,
+ "learning_rate": 0.0029313809669064377,
+ "loss": 5.946531295776367,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.03314858302474022,
+ "learning_rate": 0.0029309798927100532,
+ "loss": 5.909746170043945,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.034118980169296265,
+ "learning_rate": 0.0029305776776842453,
+ "loss": 5.903947830200195,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.03401247411966324,
+ "learning_rate": 0.0029301743221530705,
+ "loss": 5.936162948608398,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.043827857822179794,
+ "learning_rate": 0.0029297698264415047,
+ "loss": 5.953949451446533,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.06281330436468124,
+ "learning_rate": 0.002929364190875444,
+ "loss": 5.988225936889648,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.07216024398803711,
+ "learning_rate": 0.0029289574157817014,
+ "loss": 6.187741279602051,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.055205658078193665,
+ "learning_rate": 0.002928549501488008,
+ "loss": 6.177831649780273,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.04834286496043205,
+ "learning_rate": 0.0029281404483230133,
+ "loss": 6.209604263305664,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.0447777584195137,
+ "learning_rate": 0.0029277302566162837,
+ "loss": 6.146507263183594,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.05729430541396141,
+ "learning_rate": 0.0029273189266983035,
+ "loss": 6.280686378479004,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.0410807766020298,
+ "learning_rate": 0.002926906458900473,
+ "loss": 6.246824264526367,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.044607777148485184,
+ "learning_rate": 0.0029264928535551104,
+ "loss": 6.19956111907959,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.05141043663024902,
+ "learning_rate": 0.0029260781109954515,
+ "loss": 6.118578910827637,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.029248764738440514,
+ "learning_rate": 0.0029256622315556448,
+ "loss": 6.155360221862793,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.028403814882040024,
+ "learning_rate": 0.0029252452155707582,
+ "loss": 6.097352027893066,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.025872666388750076,
+ "learning_rate": 0.002924827063376773,
+ "loss": 6.072485446929932,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.018447555601596832,
+ "learning_rate": 0.0029244077753105874,
+ "loss": 6.077045440673828,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.023166058585047722,
+ "learning_rate": 0.0029239873517100147,
+ "loss": 6.040050983428955,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.019579261541366577,
+ "learning_rate": 0.002923565792913782,
+ "loss": 6.039424896240234,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.018369073048233986,
+ "learning_rate": 0.002923143099261531,
+ "loss": 6.021180152893066,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.014188461937010288,
+ "learning_rate": 0.0029227192710938195,
+ "loss": 5.9668049812316895,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.019567398354411125,
+ "learning_rate": 0.0029222943087521182,
+ "loss": 5.989189147949219,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.014190621674060822,
+ "learning_rate": 0.0029218682125788102,
+ "loss": 5.942726135253906,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01375526562333107,
+ "learning_rate": 0.002921440982917195,
+ "loss": 5.942777633666992,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.014203300699591637,
+ "learning_rate": 0.0029210126201114836,
+ "loss": 5.927928924560547,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.014231069944798946,
+ "learning_rate": 0.0029205831245067994,
+ "loss": 5.923510551452637,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.012771515175700188,
+ "learning_rate": 0.0029201524964491796,
+ "loss": 5.908803939819336,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.013399476185441017,
+ "learning_rate": 0.0029197207362855735,
+ "loss": 5.890079021453857,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.025862423703074455,
+ "learning_rate": 0.0029192878443638427,
+ "loss": 5.919907569885254,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.011315986514091492,
+ "learning_rate": 0.0029188538210327595,
+ "loss": 5.901529788970947,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.01235947571694851,
+ "learning_rate": 0.00291841866664201,
+ "loss": 5.903595924377441,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.012175127863883972,
+ "learning_rate": 0.0029179823815421893,
+ "loss": 5.868306636810303,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01195445004850626,
+ "learning_rate": 0.0029175449660848052,
+ "loss": 5.899898529052734,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.012801481410861015,
+ "learning_rate": 0.0029171064206222745,
+ "loss": 5.890621662139893,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.011909718625247478,
+ "learning_rate": 0.002916666745507926,
+ "loss": 5.86421012878418,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01162486057728529,
+ "learning_rate": 0.0029162259410959977,
+ "loss": 5.869139671325684,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.010923407040536404,
+ "learning_rate": 0.0029157840077416374,
+ "loss": 5.869789123535156,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.011943561024963856,
+ "learning_rate": 0.002915340945800904,
+ "loss": 5.873437881469727,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.010295520536601543,
+ "learning_rate": 0.002914896755630763,
+ "loss": 5.819953918457031,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.0093373479321599,
+ "learning_rate": 0.002914451437589092,
+ "loss": 5.829481601715088,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.01182442344725132,
+ "learning_rate": 0.002914004992034674,
+ "loss": 5.855249404907227,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.00939213763922453,
+ "learning_rate": 0.0029135574193272034,
+ "loss": 5.83876371383667,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.011648556217551231,
+ "learning_rate": 0.0029131087198272813,
+ "loss": 5.824906826019287,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.011227169074118137,
+ "learning_rate": 0.002912658893896416,
+ "loss": 5.847869873046875,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.011084851808845997,
+ "learning_rate": 0.002912207941897025,
+ "loss": 5.839183807373047,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.009923825040459633,
+ "learning_rate": 0.002911755864192432,
+ "loss": 5.818821907043457,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.011300318874418736,
+ "learning_rate": 0.002911302661146867,
+ "loss": 5.819517135620117,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.009915645234286785,
+ "learning_rate": 0.0029108483331254684,
+ "loss": 5.776939392089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.008965205401182175,
+ "learning_rate": 0.00291039288049428,
+ "loss": 5.824148178100586,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.011025927029550076,
+ "learning_rate": 0.0029099363036202515,
+ "loss": 5.832704544067383,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.010143411345779896,
+ "learning_rate": 0.0029094786028712387,
+ "loss": 5.802353382110596,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.010021256282925606,
+ "learning_rate": 0.002909019778616003,
+ "loss": 5.802136421203613,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.011356706731021404,
+ "learning_rate": 0.002908559831224211,
+ "loss": 5.807962417602539,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.010525298304855824,
+ "learning_rate": 0.002908098761066433,
+ "loss": 5.787661552429199,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.011147079057991505,
+ "learning_rate": 0.0029076365685141463,
+ "loss": 5.801015377044678,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.008777810260653496,
+ "learning_rate": 0.00290717325393973,
+ "loss": 5.779858589172363,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.011423609219491482,
+ "learning_rate": 0.002906708817716469,
+ "loss": 5.777166843414307,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.008639856241643429,
+ "learning_rate": 0.00290624326021855,
+ "loss": 5.747189521789551,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.009628898464143276,
+ "learning_rate": 0.002905776581821066,
+ "loss": 5.7666916847229,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.011988377198576927,
+ "learning_rate": 0.0029053087829000097,
+ "loss": 5.777951240539551,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.01693224348127842,
+ "learning_rate": 0.00290483986383228,
+ "loss": 5.818393707275391,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.020264796912670135,
+ "learning_rate": 0.002904369824995674,
+ "loss": 5.747611045837402,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.018283747136592865,
+ "learning_rate": 0.0029038986667688964,
+ "loss": 5.789663314819336,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.825676918029785,
+ "eval_runtime": 57.6908,
+ "eval_samples_per_second": 42.329,
+ "eval_steps_per_second": 1.335,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.011447477154433727,
+ "learning_rate": 0.002903426389531549,
+ "loss": 5.724975109100342,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.012539781630039215,
+ "learning_rate": 0.002902952993664138,
+ "loss": 5.747196674346924,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.012179837562143803,
+ "learning_rate": 0.002902478479548069,
+ "loss": 5.748995780944824,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.019896456971764565,
+ "learning_rate": 0.00290200284756565,
+ "loss": 5.751996994018555,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.022261742502450943,
+ "learning_rate": 0.002901526098100089,
+ "loss": 5.7516984939575195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02683536894619465,
+ "learning_rate": 0.0029010482315354946,
+ "loss": 5.716037750244141,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.018057072535157204,
+ "learning_rate": 0.0029005692482568747,
+ "loss": 5.778608322143555,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.011646891944110394,
+ "learning_rate": 0.002900089148650138,
+ "loss": 5.7308268547058105,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02230495773255825,
+ "learning_rate": 0.002899607933102092,
+ "loss": 5.768552780151367,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.01212464738637209,
+ "learning_rate": 0.002899125602000443,
+ "loss": 5.733015060424805,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.011764907278120518,
+ "learning_rate": 0.0028986421557337973,
+ "loss": 5.748104095458984,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.01593147963285446,
+ "learning_rate": 0.002898157594691658,
+ "loss": 5.701960563659668,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.016333773732185364,
+ "learning_rate": 0.0028976719192644277,
+ "loss": 5.716808795928955,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.017547868192195892,
+ "learning_rate": 0.002897185129843406,
+ "loss": 5.724377632141113,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.01662801019847393,
+ "learning_rate": 0.002896697226820791,
+ "loss": 5.737432956695557,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.015781691297888756,
+ "learning_rate": 0.002896208210589676,
+ "loss": 5.7376203536987305,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.014363214373588562,
+ "learning_rate": 0.0028957180815440534,
+ "loss": 5.7342610359191895,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.015048054046928883,
+ "learning_rate": 0.0028952268400788115,
+ "loss": 5.715073585510254,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.013753226026892662,
+ "learning_rate": 0.002894734486589734,
+ "loss": 5.73264217376709,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.015562563203275204,
+ "learning_rate": 0.002894241021473502,
+ "loss": 5.7072038650512695,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.01520819216966629,
+ "learning_rate": 0.0028937464451276916,
+ "loss": 5.683300018310547,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.013024724088609219,
+ "learning_rate": 0.002893250757950773,
+ "loss": 5.726273536682129,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.012760856188833714,
+ "learning_rate": 0.0028927539603421128,
+ "loss": 5.7249064445495605,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.013542579486966133,
+ "learning_rate": 0.0028922560527019723,
+ "loss": 5.70849084854126,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.014801754616200924,
+ "learning_rate": 0.002891757035431506,
+ "loss": 5.725185394287109,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.013000023551285267,
+ "learning_rate": 0.0028912569089327637,
+ "loss": 5.705799102783203,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.012919292785227299,
+ "learning_rate": 0.002890755673608688,
+ "loss": 5.674324989318848,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.015904201194643974,
+ "learning_rate": 0.002890253329863115,
+ "loss": 5.680788993835449,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.015467721968889236,
+ "learning_rate": 0.0028897498781007744,
+ "loss": 5.707531929016113,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.015157260000705719,
+ "learning_rate": 0.002889245318727288,
+ "loss": 5.689894676208496,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.014204306527972221,
+ "learning_rate": 0.00288873965214917,
+ "loss": 5.713001251220703,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.022045845165848732,
+ "learning_rate": 0.0028882328787738273,
+ "loss": 5.657631874084473,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.013380175456404686,
+ "learning_rate": 0.0028877249990095573,
+ "loss": 5.702192306518555,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.016698408871889114,
+ "learning_rate": 0.0028872160132655503,
+ "loss": 5.695206642150879,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.020001472905278206,
+ "learning_rate": 0.0028867059219518866,
+ "loss": 5.731742858886719,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.027350103482604027,
+ "learning_rate": 0.0028861947254795375,
+ "loss": 5.712314605712891,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.021862946450710297,
+ "learning_rate": 0.002885682424260365,
+ "loss": 5.697043418884277,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.019216977059841156,
+ "learning_rate": 0.0028851690187071204,
+ "loss": 5.698813438415527,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.02131095714867115,
+ "learning_rate": 0.002884654509233446,
+ "loss": 5.6918864250183105,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.016583887860178947,
+ "learning_rate": 0.0028841388962538724,
+ "loss": 5.711456298828125,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.016998684033751488,
+ "learning_rate": 0.0028836221801838194,
+ "loss": 5.675747871398926,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.01705755852162838,
+ "learning_rate": 0.0028831043614395966,
+ "loss": 5.697177410125732,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.014604352414608002,
+ "learning_rate": 0.0028825854404384,
+ "loss": 5.681950092315674,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.013561096973717213,
+ "learning_rate": 0.0028820654175983162,
+ "loss": 5.67357063293457,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.021149173378944397,
+ "learning_rate": 0.0028815442933383174,
+ "loss": 5.709477424621582,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.01431041955947876,
+ "learning_rate": 0.0028810220680782644,
+ "loss": 5.7012200355529785,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.01429709978401661,
+ "learning_rate": 0.0028804987422389045,
+ "loss": 5.689794540405273,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.032025713473558426,
+ "learning_rate": 0.0028799743162418712,
+ "loss": 5.6713762283325195,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.013168714940547943,
+ "learning_rate": 0.002879448790509686,
+ "loss": 5.685478210449219,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.014088666997849941,
+ "learning_rate": 0.0028789221654657556,
+ "loss": 5.645514965057373,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.013320053927600384,
+ "learning_rate": 0.002878394441534372,
+ "loss": 5.678678512573242,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.022180087864398956,
+ "learning_rate": 0.002877865619140712,
+ "loss": 5.660525321960449,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.01271889265626669,
+ "learning_rate": 0.002877335698710839,
+ "loss": 5.652164459228516,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.011617431417107582,
+ "learning_rate": 0.002876804680671701,
+ "loss": 5.6389946937561035,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.012040726840496063,
+ "learning_rate": 0.0028762725654511282,
+ "loss": 5.628744602203369,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.013148069381713867,
+ "learning_rate": 0.0028757393534778377,
+ "loss": 5.69610595703125,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.01826276071369648,
+ "learning_rate": 0.0028752050451814276,
+ "loss": 5.634238243103027,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.020393460988998413,
+ "learning_rate": 0.0028746696409923805,
+ "loss": 5.656968593597412,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.017072148621082306,
+ "learning_rate": 0.0028741331413420627,
+ "loss": 5.668219089508057,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.01694098301231861,
+ "learning_rate": 0.0028735955466627217,
+ "loss": 5.670073509216309,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.022760670632123947,
+ "learning_rate": 0.002873056857387488,
+ "loss": 5.649288177490234,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.03613442927598953,
+ "learning_rate": 0.0028725170739503734,
+ "loss": 5.723211288452148,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.04232329502701759,
+ "learning_rate": 0.002871976196786272,
+ "loss": 5.757920742034912,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.044210851192474365,
+ "learning_rate": 0.002871434226330959,
+ "loss": 5.702426910400391,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.044886402785778046,
+ "learning_rate": 0.0028708911630210905,
+ "loss": 5.742881774902344,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.022805195301771164,
+ "learning_rate": 0.002870347007294202,
+ "loss": 5.716353416442871,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02219395525753498,
+ "learning_rate": 0.00286980175958871,
+ "loss": 5.68503475189209,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.019240057095885277,
+ "learning_rate": 0.002869255420343911,
+ "loss": 5.700656890869141,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.01957295648753643,
+ "learning_rate": 0.0028687079899999807,
+ "loss": 5.70595645904541,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016326555982232094,
+ "learning_rate": 0.0028681594689979736,
+ "loss": 5.697229385375977,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015815885737538338,
+ "learning_rate": 0.0028676098577798234,
+ "loss": 5.7020792961120605,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.015865175053477287,
+ "learning_rate": 0.002867059156788342,
+ "loss": 5.648272514343262,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.7417449951171875,
+ "eval_runtime": 51.2263,
+ "eval_samples_per_second": 47.671,
+ "eval_steps_per_second": 1.503,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.016063183546066284,
+ "learning_rate": 0.0028665073664672185,
+ "loss": 5.65411376953125,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.014323881827294827,
+ "learning_rate": 0.0028659544872610214,
+ "loss": 5.629649639129639,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.014005238190293312,
+ "learning_rate": 0.002865400519615195,
+ "loss": 5.655135631561279,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.01385372132062912,
+ "learning_rate": 0.0028648454639760615,
+ "loss": 5.632213115692139,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.014073756523430347,
+ "learning_rate": 0.0028642893207908187,
+ "loss": 5.626431465148926,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.019551943987607956,
+ "learning_rate": 0.0028637320905075406,
+ "loss": 5.595447063446045,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.017571093514561653,
+ "learning_rate": 0.002863173773575179,
+ "loss": 5.661581039428711,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.01691303588449955,
+ "learning_rate": 0.0028626143704435585,
+ "loss": 5.641129493713379,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.01584242656826973,
+ "learning_rate": 0.0028620538815633807,
+ "loss": 5.640626907348633,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.014592235907912254,
+ "learning_rate": 0.0028614923073862213,
+ "loss": 5.611204147338867,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.016636405140161514,
+ "learning_rate": 0.0028609296483645304,
+ "loss": 5.614716053009033,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.01370833721011877,
+ "learning_rate": 0.002860365904951632,
+ "loss": 5.633011817932129,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.013218311592936516,
+ "learning_rate": 0.002859801077601725,
+ "loss": 5.588958740234375,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.016971616074442863,
+ "learning_rate": 0.0028592351667698786,
+ "loss": 5.635064125061035,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.015524905174970627,
+ "learning_rate": 0.0028586681729120393,
+ "loss": 5.611454963684082,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.014769157394766808,
+ "learning_rate": 0.0028581000964850215,
+ "loss": 5.62175178527832,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.014313343912363052,
+ "learning_rate": 0.002857530937946516,
+ "loss": 5.603628158569336,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.016748322173953056,
+ "learning_rate": 0.002856960697755083,
+ "loss": 5.617765426635742,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.013503300957381725,
+ "learning_rate": 0.0028563893763701538,
+ "loss": 5.636135578155518,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.012275655753910542,
+ "learning_rate": 0.002855816974252033,
+ "loss": 5.552533149719238,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.01288773212581873,
+ "learning_rate": 0.002855243491861894,
+ "loss": 5.620527267456055,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.013687008060514927,
+ "learning_rate": 0.002854668929661781,
+ "loss": 5.618595123291016,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.015678439289331436,
+ "learning_rate": 0.0028540932881146092,
+ "loss": 5.604031562805176,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.013567988760769367,
+ "learning_rate": 0.002853516567684162,
+ "loss": 5.549149513244629,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0157297495752573,
+ "learning_rate": 0.002852938768835092,
+ "loss": 5.618505477905273,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.011867019347846508,
+ "learning_rate": 0.0028523598920329233,
+ "loss": 5.579049110412598,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.01398563478142023,
+ "learning_rate": 0.0028517799377440443,
+ "loss": 5.597678184509277,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.015525273978710175,
+ "learning_rate": 0.0028511989064357148,
+ "loss": 5.601663589477539,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.01588466763496399,
+ "learning_rate": 0.002850616798576061,
+ "loss": 5.5650434494018555,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.014276969246566296,
+ "learning_rate": 0.0028500336146340772,
+ "loss": 5.607125282287598,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.013253689743578434,
+ "learning_rate": 0.002849449355079623,
+ "loss": 5.6299824714660645,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.01176661066710949,
+ "learning_rate": 0.0028488640203834263,
+ "loss": 5.560839653015137,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.011712565086781979,
+ "learning_rate": 0.0028482776110170813,
+ "loss": 5.618312835693359,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.013515647500753403,
+ "learning_rate": 0.002847690127453047,
+ "loss": 5.566764831542969,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.020177524536848068,
+ "learning_rate": 0.0028471015701646485,
+ "loss": 5.554571628570557,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.029197178781032562,
+ "learning_rate": 0.0028465119396260757,
+ "loss": 5.6097259521484375,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.027947140857577324,
+ "learning_rate": 0.0028459212363123824,
+ "loss": 5.65598201751709,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.021089809015393257,
+ "learning_rate": 0.0028453294606994887,
+ "loss": 5.67195987701416,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.019936271011829376,
+ "learning_rate": 0.0028447366132641773,
+ "loss": 5.609482765197754,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.0176444910466671,
+ "learning_rate": 0.002844142694484094,
+ "loss": 5.60432767868042,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.023154428228735924,
+ "learning_rate": 0.0028435477048377495,
+ "loss": 5.616270065307617,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.018391480669379234,
+ "learning_rate": 0.002842951644804515,
+ "loss": 5.600588321685791,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.013716557994484901,
+ "learning_rate": 0.002842354514864627,
+ "loss": 5.595095634460449,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.014372860081493855,
+ "learning_rate": 0.002841756315499181,
+ "loss": 5.584517478942871,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.014930845238268375,
+ "learning_rate": 0.002841157047190135,
+ "loss": 5.587005138397217,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.012631502002477646,
+ "learning_rate": 0.0028405567104203095,
+ "loss": 5.594094276428223,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.014677002094686031,
+ "learning_rate": 0.0028399553056733854,
+ "loss": 5.569582939147949,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.015558961778879166,
+ "learning_rate": 0.0028393528334339025,
+ "loss": 5.596535682678223,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.014322095550596714,
+ "learning_rate": 0.0028387492941872624,
+ "loss": 5.570324420928955,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.012913414277136326,
+ "learning_rate": 0.002838144688419726,
+ "loss": 5.573002338409424,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.01251310296356678,
+ "learning_rate": 0.0028375390166184137,
+ "loss": 5.550327777862549,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.013847017660737038,
+ "learning_rate": 0.0028369322792713026,
+ "loss": 5.54567813873291,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.013789358548820019,
+ "learning_rate": 0.0028363244768672313,
+ "loss": 5.571495056152344,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.01236631814390421,
+ "learning_rate": 0.0028357156098958954,
+ "loss": 5.569730758666992,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.013210169039666653,
+ "learning_rate": 0.002835105678847847,
+ "loss": 5.53984260559082,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.01604137383401394,
+ "learning_rate": 0.0028344946842144975,
+ "loss": 5.530598163604736,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.01708815060555935,
+ "learning_rate": 0.002833882626488114,
+ "loss": 5.529256820678711,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.017222166061401367,
+ "learning_rate": 0.0028332695061618197,
+ "loss": 5.564949035644531,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.01803218014538288,
+ "learning_rate": 0.002832655323729595,
+ "loss": 5.560049057006836,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.01658225618302822,
+ "learning_rate": 0.002832040079686276,
+ "loss": 5.553183555603027,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.016834143549203873,
+ "learning_rate": 0.002831423774527554,
+ "loss": 5.573678970336914,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.01856408454477787,
+ "learning_rate": 0.002830806408749974,
+ "loss": 5.539497375488281,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.026909375563263893,
+ "learning_rate": 0.002830187982850936,
+ "loss": 5.551433563232422,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.016046348959207535,
+ "learning_rate": 0.002829568497328696,
+ "loss": 5.566728591918945,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.018582617864012718,
+ "learning_rate": 0.0028289479526823614,
+ "loss": 5.543226718902588,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.0211537666618824,
+ "learning_rate": 0.0028283263494118943,
+ "loss": 5.560711860656738,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.013894101604819298,
+ "learning_rate": 0.0028277036880181095,
+ "loss": 5.528890132904053,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.017581885680556297,
+ "learning_rate": 0.0028270799690026725,
+ "loss": 5.57473087310791,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.01622696965932846,
+ "learning_rate": 0.002826455192868104,
+ "loss": 5.578961372375488,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.01966337114572525,
+ "learning_rate": 0.0028258293601177743,
+ "loss": 5.513942718505859,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021581366658210754,
+ "learning_rate": 0.002825202471255906,
+ "loss": 5.573669910430908,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.020468758419156075,
+ "learning_rate": 0.002824574526787571,
+ "loss": 5.516103744506836,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.626922130584717,
+ "eval_runtime": 54.2404,
+ "eval_samples_per_second": 45.022,
+ "eval_steps_per_second": 1.42,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.014597049914300442,
+ "learning_rate": 0.0028239455272186936,
+ "loss": 5.506453514099121,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.020277054980397224,
+ "learning_rate": 0.002823315473056048,
+ "loss": 5.535350799560547,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.019103925675153732,
+ "learning_rate": 0.0028226843648072562,
+ "loss": 5.522078037261963,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.07158710062503815,
+ "learning_rate": 0.0028220522029807913,
+ "loss": 5.531142234802246,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01958007924258709,
+ "learning_rate": 0.0028214189880859746,
+ "loss": 5.484055042266846,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.0185184795409441,
+ "learning_rate": 0.002820784720632976,
+ "loss": 5.511264801025391,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01976991444826126,
+ "learning_rate": 0.0028201494011328133,
+ "loss": 5.51821231842041,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.02188604325056076,
+ "learning_rate": 0.002819513030097352,
+ "loss": 5.483095169067383,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.019405223429203033,
+ "learning_rate": 0.002818875608039305,
+ "loss": 5.539701461791992,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.016907615587115288,
+ "learning_rate": 0.0028182371354722306,
+ "loss": 5.518251419067383,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.02160538360476494,
+ "learning_rate": 0.0028175976129105364,
+ "loss": 5.509786605834961,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.023125721141695976,
+ "learning_rate": 0.002816957040869473,
+ "loss": 5.535561561584473,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.022233380004763603,
+ "learning_rate": 0.0028163154198651372,
+ "loss": 5.553831100463867,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.017227552831172943,
+ "learning_rate": 0.002815672750414473,
+ "loss": 5.547080993652344,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.017122473567724228,
+ "learning_rate": 0.0028150290330352665,
+ "loss": 5.493778228759766,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.020641913637518883,
+ "learning_rate": 0.0028143842682461498,
+ "loss": 5.536645889282227,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.017383966594934464,
+ "learning_rate": 0.0028137384565665967,
+ "loss": 5.506233215332031,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.017969606444239616,
+ "learning_rate": 0.002813091598516928,
+ "loss": 5.524741172790527,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.01936613954603672,
+ "learning_rate": 0.0028124436946183047,
+ "loss": 5.460989952087402,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.017492495477199554,
+ "learning_rate": 0.0028117947453927313,
+ "loss": 5.487619876861572,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.018786396831274033,
+ "learning_rate": 0.0028111447513630545,
+ "loss": 5.516829013824463,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.020149117335677147,
+ "learning_rate": 0.0028104937130529624,
+ "loss": 5.460541725158691,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.024668103083968163,
+ "learning_rate": 0.0028098416309869857,
+ "loss": 5.497687339782715,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.015888608992099762,
+ "learning_rate": 0.002809188505690494,
+ "loss": 5.494459629058838,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.019327623769640923,
+ "learning_rate": 0.0028085343376896997,
+ "loss": 5.486351490020752,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.026855193078517914,
+ "learning_rate": 0.0028078791275116532,
+ "loss": 5.509825706481934,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.026340370997786522,
+ "learning_rate": 0.002807222875684246,
+ "loss": 5.527161598205566,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.02200835384428501,
+ "learning_rate": 0.002806565582736208,
+ "loss": 5.507146835327148,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.020420534536242485,
+ "learning_rate": 0.0028059072491971085,
+ "loss": 5.5162506103515625,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02493203990161419,
+ "learning_rate": 0.0028052478755973545,
+ "loss": 5.5252556800842285,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.021293973550200462,
+ "learning_rate": 0.0028045874624681925,
+ "loss": 5.523589134216309,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.016593236476182938,
+ "learning_rate": 0.0028039260103417045,
+ "loss": 5.500586032867432,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.018857723101973534,
+ "learning_rate": 0.002803263519750811,
+ "loss": 5.52933406829834,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.015930436551570892,
+ "learning_rate": 0.0028025999912292677,
+ "loss": 5.48813533782959,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.019427936524152756,
+ "learning_rate": 0.002801935425311669,
+ "loss": 5.46573543548584,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.018123053014278412,
+ "learning_rate": 0.0028012698225334427,
+ "loss": 5.499445915222168,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.02035410702228546,
+ "learning_rate": 0.002800603183430853,
+ "loss": 5.513540267944336,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.025838350877165794,
+ "learning_rate": 0.0027999355085409996,
+ "loss": 5.565637588500977,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.029621612280607224,
+ "learning_rate": 0.002799266798401816,
+ "loss": 5.5619215965271,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.028948379680514336,
+ "learning_rate": 0.0027985970535520685,
+ "loss": 5.551331520080566,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.022135930135846138,
+ "learning_rate": 0.0027979262745313604,
+ "loss": 5.527722358703613,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.028884069994091988,
+ "learning_rate": 0.0027972544618801256,
+ "loss": 5.554599761962891,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.02499174140393734,
+ "learning_rate": 0.00279658161613963,
+ "loss": 5.543351173400879,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.021131042391061783,
+ "learning_rate": 0.0027959077378519755,
+ "loss": 5.496039390563965,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.02957489714026451,
+ "learning_rate": 0.002795232827560092,
+ "loss": 5.552300453186035,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.022832119837403297,
+ "learning_rate": 0.0027945568858077433,
+ "loss": 5.488651275634766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.02074701525270939,
+ "learning_rate": 0.002793879913139523,
+ "loss": 5.491244792938232,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.019919803366065025,
+ "learning_rate": 0.002793201910100856,
+ "loss": 5.483855247497559,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.01858169585466385,
+ "learning_rate": 0.002792522877237998,
+ "loss": 5.50787353515625,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02085857465863228,
+ "learning_rate": 0.002791842815098032,
+ "loss": 5.512872695922852,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.01407450158149004,
+ "learning_rate": 0.002791161724228872,
+ "loss": 5.472590446472168,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.01672944612801075,
+ "learning_rate": 0.002790479605179262,
+ "loss": 5.469183921813965,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.016565700992941856,
+ "learning_rate": 0.0027897964584987713,
+ "loss": 5.493677139282227,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.017017360776662827,
+ "learning_rate": 0.0027891122847377996,
+ "loss": 5.464931488037109,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.019842486828565598,
+ "learning_rate": 0.002788427084447574,
+ "loss": 5.461293697357178,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.016210904344916344,
+ "learning_rate": 0.0027877408581801467,
+ "loss": 5.454716682434082,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.02064039371907711,
+ "learning_rate": 0.002787053606488398,
+ "loss": 5.466576099395752,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.018198711797595024,
+ "learning_rate": 0.0027863653299260356,
+ "loss": 5.447747230529785,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.015937626361846924,
+ "learning_rate": 0.00278567602904759,
+ "loss": 5.5055036544799805,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.01687259040772915,
+ "learning_rate": 0.002784985704408419,
+ "loss": 5.422250747680664,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.02622046135365963,
+ "learning_rate": 0.002784294356564705,
+ "loss": 5.466894626617432,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.018672186881303787,
+ "learning_rate": 0.0027836019860734535,
+ "loss": 5.48300838470459,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.02166241593658924,
+ "learning_rate": 0.0027829085934924958,
+ "loss": 5.477658271789551,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018426023423671722,
+ "learning_rate": 0.0027822141793804856,
+ "loss": 5.487186431884766,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.022229891270399094,
+ "learning_rate": 0.0027815187442968995,
+ "loss": 5.488579273223877,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.026930972933769226,
+ "learning_rate": 0.002780822288802037,
+ "loss": 5.492000102996826,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.02733265422284603,
+ "learning_rate": 0.00278012481345702,
+ "loss": 5.4663801193237305,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.02993253618478775,
+ "learning_rate": 0.0027794263188237922,
+ "loss": 5.543368816375732,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03152153640985489,
+ "learning_rate": 0.002778726805465118,
+ "loss": 5.565644264221191,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.031395044177770615,
+ "learning_rate": 0.002778026273944581,
+ "loss": 5.553359508514404,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.026603225618600845,
+ "learning_rate": 0.002777324724826589,
+ "loss": 5.552347183227539,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.024152139201760292,
+ "learning_rate": 0.002776622158676366,
+ "loss": 5.493257522583008,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.602636814117432,
+ "eval_runtime": 54.6199,
+ "eval_samples_per_second": 44.709,
+ "eval_steps_per_second": 1.41,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.02089761756360531,
+ "learning_rate": 0.0027759185760599576,
+ "loss": 5.461747646331787,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.018181778490543365,
+ "learning_rate": 0.0027752139775442278,
+ "loss": 5.492741584777832,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.020166238769888878,
+ "learning_rate": 0.0027745083636968575,
+ "loss": 5.476003646850586,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.015491694211959839,
+ "learning_rate": 0.0027738017350863488,
+ "loss": 5.425764560699463,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.022291697561740875,
+ "learning_rate": 0.0027730940922820182,
+ "loss": 5.45866584777832,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.018074683845043182,
+ "learning_rate": 0.002772385435854002,
+ "loss": 5.465143203735352,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.016872268170118332,
+ "learning_rate": 0.002771675766373251,
+ "loss": 5.446003437042236,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.023103339597582817,
+ "learning_rate": 0.0027709650844115336,
+ "loss": 5.411036491394043,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.027957158163189888,
+ "learning_rate": 0.0027702533905414324,
+ "loss": 5.452442169189453,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.019935399293899536,
+ "learning_rate": 0.002769540685336348,
+ "loss": 5.412145614624023,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.014351371675729752,
+ "learning_rate": 0.0027688269693704926,
+ "loss": 5.434524059295654,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.0152254244312644,
+ "learning_rate": 0.0027681122432188948,
+ "loss": 5.428544998168945,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.015071330592036247,
+ "learning_rate": 0.0027673965074573965,
+ "loss": 5.459929466247559,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.023415958508849144,
+ "learning_rate": 0.0027666797626626537,
+ "loss": 5.447159767150879,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.017228802666068077,
+ "learning_rate": 0.0027659620094121337,
+ "loss": 5.442273139953613,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.014184504747390747,
+ "learning_rate": 0.002765243248284118,
+ "loss": 5.448305130004883,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.024384526535868645,
+ "learning_rate": 0.0027645234798576988,
+ "loss": 5.4076337814331055,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.043152231723070145,
+ "learning_rate": 0.0027638027047127813,
+ "loss": 5.45344352722168,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.02889758162200451,
+ "learning_rate": 0.002763080923430081,
+ "loss": 5.489194393157959,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.02686968818306923,
+ "learning_rate": 0.002762358136591123,
+ "loss": 5.493659496307373,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.020467868074774742,
+ "learning_rate": 0.002761634344778245,
+ "loss": 5.410489559173584,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.019759224727749825,
+ "learning_rate": 0.002760909548574592,
+ "loss": 5.439385890960693,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.0233843382447958,
+ "learning_rate": 0.0027601837485641194,
+ "loss": 5.445036888122559,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.031753458082675934,
+ "learning_rate": 0.0027594569453315916,
+ "loss": 5.503895282745361,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.02089349366724491,
+ "learning_rate": 0.002758729139462581,
+ "loss": 5.45689582824707,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.03339357674121857,
+ "learning_rate": 0.002758000331543466,
+ "loss": 5.522361755371094,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019328996539115906,
+ "learning_rate": 0.002757270522161436,
+ "loss": 5.441373825073242,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019863856956362724,
+ "learning_rate": 0.002756539711904484,
+ "loss": 5.4633283615112305,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.02141754701733589,
+ "learning_rate": 0.0027558079013614116,
+ "loss": 5.453215599060059,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016591915860772133,
+ "learning_rate": 0.002755075091121824,
+ "loss": 5.463696479797363,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.01955028623342514,
+ "learning_rate": 0.0027543412817761345,
+ "loss": 5.411355018615723,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.016791662201285362,
+ "learning_rate": 0.0027536064739155596,
+ "loss": 5.459614276885986,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.019661981612443924,
+ "learning_rate": 0.00275287066813212,
+ "loss": 5.440699577331543,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.01927965320646763,
+ "learning_rate": 0.002752133865018642,
+ "loss": 5.409865856170654,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.016483580693602562,
+ "learning_rate": 0.002751396065168754,
+ "loss": 5.390999794006348,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016264773905277252,
+ "learning_rate": 0.0027506572691768888,
+ "loss": 5.418543815612793,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.01958460919559002,
+ "learning_rate": 0.00274991747763828,
+ "loss": 5.403289794921875,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.01861695572733879,
+ "learning_rate": 0.002749176691148965,
+ "loss": 5.431911468505859,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.029594263061881065,
+ "learning_rate": 0.002748434910305782,
+ "loss": 5.399242401123047,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.028574613854289055,
+ "learning_rate": 0.0027476921357063697,
+ "loss": 5.440446853637695,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.028612585738301277,
+ "learning_rate": 0.0027469483679491693,
+ "loss": 5.410346031188965,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.020704898983240128,
+ "learning_rate": 0.00274620360763342,
+ "loss": 5.44694709777832,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.020542364567518234,
+ "learning_rate": 0.002745457855359163,
+ "loss": 5.402653694152832,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.015762662515044212,
+ "learning_rate": 0.0027447111117272363,
+ "loss": 5.396594047546387,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022096984088420868,
+ "learning_rate": 0.002743963377339278,
+ "loss": 5.398744106292725,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.018791764974594116,
+ "learning_rate": 0.002743214652797724,
+ "loss": 5.397758483886719,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01642526686191559,
+ "learning_rate": 0.0027424649387058093,
+ "loss": 5.411459922790527,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.017931703478097916,
+ "learning_rate": 0.0027417142356675637,
+ "loss": 5.390264511108398,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018819602206349373,
+ "learning_rate": 0.0027409625442878153,
+ "loss": 5.407539367675781,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.019389411434531212,
+ "learning_rate": 0.0027402098651721886,
+ "loss": 5.386474609375,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.013645983301103115,
+ "learning_rate": 0.002739456198927104,
+ "loss": 5.373845100402832,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.013846895657479763,
+ "learning_rate": 0.002738701546159776,
+ "loss": 5.387155055999756,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013553125783801079,
+ "learning_rate": 0.002737945907478215,
+ "loss": 5.377801895141602,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.012771612964570522,
+ "learning_rate": 0.002737189283491225,
+ "loss": 5.344066143035889,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.020465301349759102,
+ "learning_rate": 0.002736431674808405,
+ "loss": 5.382771968841553,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.014798106625676155,
+ "learning_rate": 0.0027356730820401465,
+ "loss": 5.370522499084473,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.01566634140908718,
+ "learning_rate": 0.0027349135057976332,
+ "loss": 5.351980209350586,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.017485102638602257,
+ "learning_rate": 0.002734152946692843,
+ "loss": 5.355443477630615,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.016646096482872963,
+ "learning_rate": 0.0027333914053385444,
+ "loss": 5.34829044342041,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.017293352633714676,
+ "learning_rate": 0.0027326288823482964,
+ "loss": 5.3519487380981445,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.018873002380132675,
+ "learning_rate": 0.002731865378336452,
+ "loss": 5.348442077636719,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.015504784882068634,
+ "learning_rate": 0.002731100893918151,
+ "loss": 5.341490268707275,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.014270870015025139,
+ "learning_rate": 0.0027303354297093254,
+ "loss": 5.359251022338867,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.018820680677890778,
+ "learning_rate": 0.002729568986326695,
+ "loss": 5.390280723571777,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.021008187904953957,
+ "learning_rate": 0.0027288015643877708,
+ "loss": 5.338380336761475,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.016266655176877975,
+ "learning_rate": 0.0027280331645108498,
+ "loss": 5.347043037414551,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.01831045374274254,
+ "learning_rate": 0.0027272637873150185,
+ "loss": 5.326537132263184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02073794975876808,
+ "learning_rate": 0.002726493433420149,
+ "loss": 5.3607025146484375,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.022534754127264023,
+ "learning_rate": 0.0027257221034469033,
+ "loss": 5.353292465209961,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.021627362817525864,
+ "learning_rate": 0.002724949798016727,
+ "loss": 5.373323917388916,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.018226243555545807,
+ "learning_rate": 0.0027241765177518523,
+ "loss": 5.344578266143799,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.02032875455915928,
+ "learning_rate": 0.002723402263275298,
+ "loss": 5.329159259796143,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.44969367980957,
+ "eval_runtime": 51.2294,
+ "eval_samples_per_second": 47.668,
+ "eval_steps_per_second": 1.503,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.01613701693713665,
+ "learning_rate": 0.0027226270352108664,
+ "loss": 5.3024678230285645,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01940973661839962,
+ "learning_rate": 0.0027218508341831445,
+ "loss": 5.270181179046631,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.015840493142604828,
+ "learning_rate": 0.002721073660817504,
+ "loss": 5.261706352233887,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.017390675842761993,
+ "learning_rate": 0.0027202955157401,
+ "loss": 5.262207984924316,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.015184012241661549,
+ "learning_rate": 0.0027195163995778686,
+ "loss": 5.297155380249023,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.013417073525488377,
+ "learning_rate": 0.00271873631295853,
+ "loss": 5.279838562011719,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.013024317100644112,
+ "learning_rate": 0.0027179552565105864,
+ "loss": 5.265446186065674,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.01426338218152523,
+ "learning_rate": 0.002717173230863321,
+ "loss": 5.284180164337158,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.019604040309786797,
+ "learning_rate": 0.002716390236646797,
+ "loss": 5.286281108856201,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.019039476290345192,
+ "learning_rate": 0.0027156062744918587,
+ "loss": 5.290433406829834,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.019284334033727646,
+ "learning_rate": 0.0027148213450301313,
+ "loss": 5.262433052062988,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.0145154669880867,
+ "learning_rate": 0.0027140354488940175,
+ "loss": 5.24387264251709,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.014776417054235935,
+ "learning_rate": 0.0027132485867166992,
+ "loss": 5.282435417175293,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.012711729854345322,
+ "learning_rate": 0.0027124607591321377,
+ "loss": 5.246735572814941,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.013394835405051708,
+ "learning_rate": 0.0027116719667750716,
+ "loss": 5.270739555358887,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.013468850404024124,
+ "learning_rate": 0.002710882210281016,
+ "loss": 5.27161169052124,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.030262840911746025,
+ "learning_rate": 0.002710091490286264,
+ "loss": 5.256673812866211,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.013560288585722446,
+ "learning_rate": 0.0027092998074278846,
+ "loss": 5.253674507141113,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014592813327908516,
+ "learning_rate": 0.002708507162343721,
+ "loss": 5.25804328918457,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014328169636428356,
+ "learning_rate": 0.0027077135556723953,
+ "loss": 5.235117435455322,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.01277366653084755,
+ "learning_rate": 0.0027069189880533,
+ "loss": 5.2053985595703125,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.012948785908520222,
+ "learning_rate": 0.0027061234601266043,
+ "loss": 5.214653015136719,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.014892828650772572,
+ "learning_rate": 0.002705326972533252,
+ "loss": 5.227547645568848,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01427230704575777,
+ "learning_rate": 0.002704529525914957,
+ "loss": 5.241065502166748,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.01650116965174675,
+ "learning_rate": 0.0027037311209142092,
+ "loss": 5.227713108062744,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.019728578627109528,
+ "learning_rate": 0.0027029317581742677,
+ "loss": 5.2624921798706055,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.023847149685025215,
+ "learning_rate": 0.0027021314383391657,
+ "loss": 5.278811454772949,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.020690469071269035,
+ "learning_rate": 0.0027013301620537056,
+ "loss": 5.238919258117676,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.021211959421634674,
+ "learning_rate": 0.0027005279299634616,
+ "loss": 5.253684997558594,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.025854922831058502,
+ "learning_rate": 0.002699724742714778,
+ "loss": 5.2538652420043945,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.0215485580265522,
+ "learning_rate": 0.0026989206009547678,
+ "loss": 5.224540710449219,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.019827792420983315,
+ "learning_rate": 0.0026981155053313137,
+ "loss": 5.220674514770508,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.017689093947410583,
+ "learning_rate": 0.002697309456493066,
+ "loss": 5.228511810302734,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.022564806044101715,
+ "learning_rate": 0.002696502455089445,
+ "loss": 5.295933723449707,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.020070280879735947,
+ "learning_rate": 0.0026956945017706357,
+ "loss": 5.230920791625977,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.025744661688804626,
+ "learning_rate": 0.002694885597187593,
+ "loss": 5.296453952789307,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.021263007074594498,
+ "learning_rate": 0.0026940757419920357,
+ "loss": 5.220481872558594,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.017577245831489563,
+ "learning_rate": 0.0026932649368364495,
+ "loss": 5.231488227844238,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.015562832355499268,
+ "learning_rate": 0.002692453182374086,
+ "loss": 5.175224304199219,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.014115403406322002,
+ "learning_rate": 0.0026916404792589606,
+ "loss": 5.228387832641602,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.012863830663263798,
+ "learning_rate": 0.002690826828145854,
+ "loss": 5.186727523803711,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.014151126146316528,
+ "learning_rate": 0.002690012229690309,
+ "loss": 5.1694817543029785,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.01316932775080204,
+ "learning_rate": 0.0026891966845486353,
+ "loss": 5.19931697845459,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.011340726166963577,
+ "learning_rate": 0.0026883801933779003,
+ "loss": 5.238619804382324,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.012616089545190334,
+ "learning_rate": 0.0026875627568359378,
+ "loss": 5.183376789093018,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.012274038046598434,
+ "learning_rate": 0.002686744375581341,
+ "loss": 5.199946403503418,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.014990398660302162,
+ "learning_rate": 0.002685925050273465,
+ "loss": 5.168451309204102,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.011399757117033005,
+ "learning_rate": 0.0026851047815724255,
+ "loss": 5.173828125,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.013638390228152275,
+ "learning_rate": 0.0026842835701390982,
+ "loss": 5.142702102661133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.014472718350589275,
+ "learning_rate": 0.0026834614166351177,
+ "loss": 5.173436164855957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.014509808272123337,
+ "learning_rate": 0.002682638321722879,
+ "loss": 5.173975467681885,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.013576583005487919,
+ "learning_rate": 0.0026818142860655352,
+ "loss": 5.1435089111328125,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.012914472259581089,
+ "learning_rate": 0.0026809893103269954,
+ "loss": 5.151381492614746,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.013803352601826191,
+ "learning_rate": 0.0026801633951719288,
+ "loss": 5.133159637451172,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.018488213419914246,
+ "learning_rate": 0.0026793365412657606,
+ "loss": 5.183903694152832,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.01843969337642193,
+ "learning_rate": 0.002678508749274671,
+ "loss": 5.1660356521606445,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.017181839793920517,
+ "learning_rate": 0.0026776800198655983,
+ "loss": 5.179847717285156,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0148160420358181,
+ "learning_rate": 0.0026768503537062333,
+ "loss": 5.149198532104492,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.013251296244561672,
+ "learning_rate": 0.0026760197514650247,
+ "loss": 5.192000389099121,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.010746927931904793,
+ "learning_rate": 0.0026751882138111717,
+ "loss": 5.138680458068848,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.012316994369029999,
+ "learning_rate": 0.002674355741414631,
+ "loss": 5.169965744018555,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.01191797386854887,
+ "learning_rate": 0.0026735223349461106,
+ "loss": 5.128267288208008,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.013525084592401981,
+ "learning_rate": 0.0026726879950770698,
+ "loss": 5.124795436859131,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.014843370765447617,
+ "learning_rate": 0.002671852722479721,
+ "loss": 5.1497087478637695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.015285412780940533,
+ "learning_rate": 0.00267101651782703,
+ "loss": 5.093118190765381,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.013555832207202911,
+ "learning_rate": 0.0026701793817927107,
+ "loss": 5.13249397277832,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.010569742880761623,
+ "learning_rate": 0.0026693413150512273,
+ "loss": 5.084077835083008,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.011483813636004925,
+ "learning_rate": 0.0026685023182777975,
+ "loss": 5.128049850463867,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.013724366202950478,
+ "learning_rate": 0.0026676623921483844,
+ "loss": 5.157539367675781,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.014337432570755482,
+ "learning_rate": 0.0026668215373397016,
+ "loss": 5.149948596954346,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.014101454056799412,
+ "learning_rate": 0.00266597975452921,
+ "loss": 5.117250442504883,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.016004368662834167,
+ "learning_rate": 0.0026651370443951197,
+ "loss": 5.143221855163574,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.249833583831787,
+ "eval_runtime": 51.5543,
+ "eval_samples_per_second": 47.368,
+ "eval_steps_per_second": 1.494,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.01297403872013092,
+ "learning_rate": 0.0026642934076163867,
+ "loss": 5.089287281036377,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.012867718935012817,
+ "learning_rate": 0.0026634488448727144,
+ "loss": 5.0784101486206055,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.014600927010178566,
+ "learning_rate": 0.0026626033568445504,
+ "loss": 5.050220489501953,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.016619572415947914,
+ "learning_rate": 0.0026617569442130905,
+ "loss": 5.079303741455078,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.015560587868094444,
+ "learning_rate": 0.0026609096076602736,
+ "loss": 5.072385311126709,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.01774057187139988,
+ "learning_rate": 0.0026600613478687833,
+ "loss": 5.122529983520508,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.019971400499343872,
+ "learning_rate": 0.0026592121655220472,
+ "loss": 5.085966110229492,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.017667256295681,
+ "learning_rate": 0.0026583620613042365,
+ "loss": 5.078432559967041,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.022369660437107086,
+ "learning_rate": 0.002657511035900264,
+ "loss": 5.0630903244018555,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.017964458093047142,
+ "learning_rate": 0.002656659089995787,
+ "loss": 5.0846076011657715,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.021769145503640175,
+ "learning_rate": 0.002655806224277201,
+ "loss": 5.081344127655029,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.02487620897591114,
+ "learning_rate": 0.0026549524394316464,
+ "loss": 5.106287956237793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.022516218945384026,
+ "learning_rate": 0.0026540977361470012,
+ "loss": 5.128098487854004,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.02099926583468914,
+ "learning_rate": 0.0026532421151118845,
+ "loss": 5.103562355041504,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.01899387687444687,
+ "learning_rate": 0.002652385577015655,
+ "loss": 5.062675476074219,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.016801996156573296,
+ "learning_rate": 0.002651528122548409,
+ "loss": 5.096002578735352,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.014009200967848301,
+ "learning_rate": 0.0026506697524009836,
+ "loss": 5.075201034545898,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.014653858728706837,
+ "learning_rate": 0.002649810467264951,
+ "loss": 5.090478897094727,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.014918150380253792,
+ "learning_rate": 0.0026489502678326226,
+ "loss": 5.070797920227051,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.01327288057655096,
+ "learning_rate": 0.0026480891547970443,
+ "loss": 5.030500411987305,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.012885140255093575,
+ "learning_rate": 0.002647227128852,
+ "loss": 5.0845537185668945,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.012848512269556522,
+ "learning_rate": 0.002646364190692009,
+ "loss": 5.079484939575195,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.01149789709597826,
+ "learning_rate": 0.002645500341012324,
+ "loss": 5.036316394805908,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.010360757820308208,
+ "learning_rate": 0.0026446355805089337,
+ "loss": 5.04129695892334,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.010426290333271027,
+ "learning_rate": 0.00264376990987856,
+ "loss": 5.051270484924316,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.01018153689801693,
+ "learning_rate": 0.002642903329818658,
+ "loss": 5.102342128753662,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.011879283003509045,
+ "learning_rate": 0.0026420358410274153,
+ "loss": 5.053501129150391,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.010418329387903214,
+ "learning_rate": 0.002641167444203752,
+ "loss": 5.030653476715088,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.012302013114094734,
+ "learning_rate": 0.0026402981400473202,
+ "loss": 5.0402069091796875,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.011392916552722454,
+ "learning_rate": 0.002639427929258502,
+ "loss": 5.024111747741699,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.013075058348476887,
+ "learning_rate": 0.0026385568125384114,
+ "loss": 5.021731376647949,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015349832363426685,
+ "learning_rate": 0.0026376847905888903,
+ "loss": 5.03192138671875,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01588309183716774,
+ "learning_rate": 0.0026368118641125113,
+ "loss": 4.966363906860352,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.014378590509295464,
+ "learning_rate": 0.0026359380338125764,
+ "loss": 5.070773124694824,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.012306655757129192,
+ "learning_rate": 0.002635063300393114,
+ "loss": 5.025704860687256,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01212004292756319,
+ "learning_rate": 0.002634187664558881,
+ "loss": 4.965104103088379,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013538849540054798,
+ "learning_rate": 0.0026333111270153624,
+ "loss": 5.029810905456543,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.012052906677126884,
+ "learning_rate": 0.0026324336884687672,
+ "loss": 5.0262908935546875,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.011368660256266594,
+ "learning_rate": 0.002631555349626033,
+ "loss": 5.001485824584961,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.010215395130217075,
+ "learning_rate": 0.0026306761111948205,
+ "loss": 5.019739151000977,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.011449878104031086,
+ "learning_rate": 0.0026297959738835167,
+ "loss": 4.9816083908081055,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.012381811626255512,
+ "learning_rate": 0.0026289149384012326,
+ "loss": 5.025032043457031,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.00974641740322113,
+ "learning_rate": 0.002628033005457802,
+ "loss": 5.000177383422852,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.009515874087810516,
+ "learning_rate": 0.0026271501757637833,
+ "loss": 4.996979713439941,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.011125455610454082,
+ "learning_rate": 0.0026262664500304555,
+ "loss": 4.996796607971191,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.013290468603372574,
+ "learning_rate": 0.0026253818289698207,
+ "loss": 5.053703308105469,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.014650283381342888,
+ "learning_rate": 0.002624496313294602,
+ "loss": 5.001321792602539,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014537827111780643,
+ "learning_rate": 0.002623609903718244,
+ "loss": 5.024584770202637,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.012908566743135452,
+ "learning_rate": 0.0026227226009549105,
+ "loss": 5.042878150939941,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.01446808222681284,
+ "learning_rate": 0.002621834405719485,
+ "loss": 5.020857810974121,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.01535153016448021,
+ "learning_rate": 0.0026209453187275708,
+ "loss": 5.0056304931640625,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01815616339445114,
+ "learning_rate": 0.0026200553406954892,
+ "loss": 5.0467376708984375,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.023135516792535782,
+ "learning_rate": 0.0026191644723402795,
+ "loss": 5.033995628356934,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.019682930782437325,
+ "learning_rate": 0.002618272714379698,
+ "loss": 5.008149147033691,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.018521904945373535,
+ "learning_rate": 0.002617380067532217,
+ "loss": 5.007076263427734,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.02118663489818573,
+ "learning_rate": 0.0026164865325170278,
+ "loss": 5.05519962310791,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024684788659214973,
+ "learning_rate": 0.0026155921100540344,
+ "loss": 5.027571201324463,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.02609819918870926,
+ "learning_rate": 0.0026146968008638567,
+ "loss": 5.10361385345459,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.025958053767681122,
+ "learning_rate": 0.00261380060566783,
+ "loss": 5.0911455154418945,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.017067518085241318,
+ "learning_rate": 0.002612903525188002,
+ "loss": 5.045783042907715,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.016510486602783203,
+ "learning_rate": 0.0026120055601471345,
+ "loss": 5.069464683532715,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015552333556115627,
+ "learning_rate": 0.0026111067112687012,
+ "loss": 5.024248123168945,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.013432685285806656,
+ "learning_rate": 0.0026102069792768886,
+ "loss": 5.004837989807129,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.012300201691687107,
+ "learning_rate": 0.0026093063648965943,
+ "loss": 5.031093597412109,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.01197744719684124,
+ "learning_rate": 0.002608404868853428,
+ "loss": 4.965237617492676,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.01100246049463749,
+ "learning_rate": 0.002607502491873708,
+ "loss": 5.04719352722168,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.010111447423696518,
+ "learning_rate": 0.0026065992346844626,
+ "loss": 4.986410617828369,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.010669790208339691,
+ "learning_rate": 0.002605695098013431,
+ "loss": 4.9825663566589355,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.010805477388203144,
+ "learning_rate": 0.0026047900825890593,
+ "loss": 4.991506576538086,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.010886074975132942,
+ "learning_rate": 0.002603884189140502,
+ "loss": 4.987531661987305,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.010735839605331421,
+ "learning_rate": 0.0026029774183976206,
+ "loss": 4.982756614685059,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.011092593893408775,
+ "learning_rate": 0.002602069771090985,
+ "loss": 4.994824409484863,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 5.13364315032959,
+ "eval_runtime": 50.4195,
+ "eval_samples_per_second": 48.434,
+ "eval_steps_per_second": 1.527,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.01005905494093895,
+ "learning_rate": 0.0026011612479518694,
+ "loss": 4.918412685394287,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.009904682636260986,
+ "learning_rate": 0.002600251849712255,
+ "loss": 4.924630165100098,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.009543103165924549,
+ "learning_rate": 0.0025993415771048273,
+ "loss": 4.940309524536133,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.011031470261514187,
+ "learning_rate": 0.0025984304308629765,
+ "loss": 4.924609184265137,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.009511022828519344,
+ "learning_rate": 0.002597518411720796,
+ "loss": 4.9102325439453125,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.00886621605604887,
+ "learning_rate": 0.0025966055204130847,
+ "loss": 4.9236907958984375,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.010057039558887482,
+ "learning_rate": 0.002595691757675341,
+ "loss": 4.908633232116699,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.01098555512726307,
+ "learning_rate": 0.0025947771242437684,
+ "loss": 4.935927391052246,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.008800056762993336,
+ "learning_rate": 0.0025938616208552694,
+ "loss": 4.873291969299316,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.009747766889631748,
+ "learning_rate": 0.0025929452482474497,
+ "loss": 4.954436302185059,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.010785684920847416,
+ "learning_rate": 0.002592028007158613,
+ "loss": 4.924927711486816,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.012620717287063599,
+ "learning_rate": 0.0025911098983277645,
+ "loss": 4.953606605529785,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.014428738504648209,
+ "learning_rate": 0.002590190922494607,
+ "loss": 4.951706886291504,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.014496048912405968,
+ "learning_rate": 0.0025892710803995444,
+ "loss": 4.941249847412109,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.0109238987788558,
+ "learning_rate": 0.0025883503727836764,
+ "loss": 4.939975261688232,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.012271072715520859,
+ "learning_rate": 0.0025874288003887996,
+ "loss": 4.880942344665527,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.016302887350320816,
+ "learning_rate": 0.0025865063639574086,
+ "loss": 4.931602478027344,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.016137251630425453,
+ "learning_rate": 0.0025855830642326945,
+ "loss": 4.918013095855713,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.014732961542904377,
+ "learning_rate": 0.002584658901958542,
+ "loss": 4.927577495574951,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.016387946903705597,
+ "learning_rate": 0.0025837338778795325,
+ "loss": 4.930819034576416,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.017146151512861252,
+ "learning_rate": 0.002582807992740941,
+ "loss": 4.928377151489258,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.016153613105416298,
+ "learning_rate": 0.0025818812472887375,
+ "loss": 4.975607395172119,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.013386047445237637,
+ "learning_rate": 0.0025809536422695825,
+ "loss": 4.9434709548950195,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.012328253127634525,
+ "learning_rate": 0.0025800251784308314,
+ "loss": 4.917585372924805,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.012000586837530136,
+ "learning_rate": 0.0025790958565205316,
+ "loss": 4.9137725830078125,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.009463997557759285,
+ "learning_rate": 0.002578165677287419,
+ "loss": 4.910362243652344,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.010591360740363598,
+ "learning_rate": 0.0025772346414809242,
+ "loss": 4.916032314300537,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.010429762303829193,
+ "learning_rate": 0.002576302749851164,
+ "loss": 4.934561729431152,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.12526583671569824,
+ "learning_rate": 0.0025753700031489486,
+ "loss": 4.897515773773193,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01368892751634121,
+ "learning_rate": 0.0025744364021257743,
+ "loss": 4.94814920425415,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.016603967174887657,
+ "learning_rate": 0.0025735019475338254,
+ "loss": 4.8996758460998535,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.06110106781125069,
+ "learning_rate": 0.0025725666401259775,
+ "loss": 4.946794509887695,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.018292423337697983,
+ "learning_rate": 0.002571630480655789,
+ "loss": 4.958992958068848,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.01852245442569256,
+ "learning_rate": 0.0025706934698775073,
+ "loss": 4.942556381225586,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.0195649154484272,
+ "learning_rate": 0.0025697556085460646,
+ "loss": 4.895773887634277,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.017820244655013084,
+ "learning_rate": 0.00256881689741708,
+ "loss": 4.915250778198242,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.016814742237329483,
+ "learning_rate": 0.002567877337246855,
+ "loss": 4.9521260261535645,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.015319781377911568,
+ "learning_rate": 0.002566936928792377,
+ "loss": 4.899402141571045,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.01607401855289936,
+ "learning_rate": 0.0025659956728113146,
+ "loss": 4.909913063049316,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.015564955770969391,
+ "learning_rate": 0.002565053570062023,
+ "loss": 4.947719573974609,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.017502037808299065,
+ "learning_rate": 0.002564110621303536,
+ "loss": 4.95107364654541,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.014791073277592659,
+ "learning_rate": 0.002563166827295571,
+ "loss": 4.8995866775512695,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.012700124643743038,
+ "learning_rate": 0.002562222188798525,
+ "loss": 4.922100067138672,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.011235272511839867,
+ "learning_rate": 0.0025612767065734763,
+ "loss": 4.884456634521484,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.013009768910706043,
+ "learning_rate": 0.002560330381382183,
+ "loss": 4.869503021240234,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.0102638378739357,
+ "learning_rate": 0.0025593832139870834,
+ "loss": 4.934551239013672,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.011081870645284653,
+ "learning_rate": 0.0025584352051512913,
+ "loss": 4.937268257141113,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.010681218467652798,
+ "learning_rate": 0.002557486355638601,
+ "loss": 4.920559883117676,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.010962890461087227,
+ "learning_rate": 0.002556536666213484,
+ "loss": 4.936082363128662,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.010790379717946053,
+ "learning_rate": 0.0025555861376410868,
+ "loss": 4.9506449699401855,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.010659157298505306,
+ "learning_rate": 0.0025546347706872346,
+ "loss": 4.872883319854736,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.012962655164301395,
+ "learning_rate": 0.0025536825661184254,
+ "loss": 4.901210784912109,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.014810293912887573,
+ "learning_rate": 0.002552729524701833,
+ "loss": 4.895804405212402,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.010993832722306252,
+ "learning_rate": 0.0025517756472053068,
+ "loss": 4.891035079956055,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.013413915410637856,
+ "learning_rate": 0.002550820934397368,
+ "loss": 4.900317192077637,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.01580147072672844,
+ "learning_rate": 0.0025498653870472117,
+ "loss": 4.883669853210449,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01864512264728546,
+ "learning_rate": 0.0025489090059247045,
+ "loss": 4.881938457489014,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017185743898153305,
+ "learning_rate": 0.0025479517918003862,
+ "loss": 4.853325843811035,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.012045156210660934,
+ "learning_rate": 0.002546993745445466,
+ "loss": 4.9152326583862305,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.013420728035271168,
+ "learning_rate": 0.0025460348676318257,
+ "loss": 4.895130157470703,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.011561610735952854,
+ "learning_rate": 0.0025450751591320152,
+ "loss": 4.890257835388184,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013164008036255836,
+ "learning_rate": 0.0025441146207192547,
+ "loss": 4.885972023010254,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01120204571634531,
+ "learning_rate": 0.002543153253167432,
+ "loss": 4.882246971130371,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.011342627927660942,
+ "learning_rate": 0.002542191057251104,
+ "loss": 4.862927436828613,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.010852213948965073,
+ "learning_rate": 0.0025412280337454942,
+ "loss": 4.879744529724121,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.012851282022893429,
+ "learning_rate": 0.0025402641834264936,
+ "loss": 4.882238388061523,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.011839655227959156,
+ "learning_rate": 0.002539299507070659,
+ "loss": 4.85658073425293,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.011315262876451015,
+ "learning_rate": 0.002538334005455213,
+ "loss": 4.9234771728515625,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.011661105789244175,
+ "learning_rate": 0.0025373676793580416,
+ "loss": 4.855696678161621,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.009712521918118,
+ "learning_rate": 0.0025364005295576975,
+ "loss": 4.89942741394043,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.011002928018569946,
+ "learning_rate": 0.0025354325568333956,
+ "loss": 4.871459007263184,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.01509513333439827,
+ "learning_rate": 0.0025344637619650135,
+ "loss": 4.834778785705566,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 5.031810283660889,
+ "eval_runtime": 51.9412,
+ "eval_samples_per_second": 47.015,
+ "eval_steps_per_second": 1.482,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.010317740961909294,
+ "learning_rate": 0.0025334941457330925,
+ "loss": 4.8169708251953125,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.00892715621739626,
+ "learning_rate": 0.0025325237089188343,
+ "loss": 4.852178573608398,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.011608307249844074,
+ "learning_rate": 0.002531552452304103,
+ "loss": 4.790487289428711,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.013100114651024342,
+ "learning_rate": 0.0025305803766714216,
+ "loss": 4.863536834716797,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.016571057960391045,
+ "learning_rate": 0.002529607482803974,
+ "loss": 4.840987205505371,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.014595622196793556,
+ "learning_rate": 0.002528633771485604,
+ "loss": 4.815027236938477,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.014603931456804276,
+ "learning_rate": 0.002527659243500812,
+ "loss": 4.853024959564209,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.019107041880488396,
+ "learning_rate": 0.0025266838996347595,
+ "loss": 4.829808235168457,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.020292190834879875,
+ "learning_rate": 0.0025257077406732613,
+ "loss": 4.775232315063477,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.014630074612796307,
+ "learning_rate": 0.0025247307674027926,
+ "loss": 4.826682090759277,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.019507668912410736,
+ "learning_rate": 0.0025237529806104823,
+ "loss": 4.843222141265869,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.023950330913066864,
+ "learning_rate": 0.002522774381084116,
+ "loss": 4.820512771606445,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.022547820582985878,
+ "learning_rate": 0.002521794969612133,
+ "loss": 4.854910850524902,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.017819277942180634,
+ "learning_rate": 0.0025208147469836283,
+ "loss": 4.825885772705078,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.017989592626690865,
+ "learning_rate": 0.002519833713988349,
+ "loss": 4.8357014656066895,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.013516640290617943,
+ "learning_rate": 0.0025188518714166956,
+ "loss": 4.83729362487793,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.015956101939082146,
+ "learning_rate": 0.002517869220059721,
+ "loss": 4.807845592498779,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.013967473991215229,
+ "learning_rate": 0.002516885760709131,
+ "loss": 4.828944206237793,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.01270931027829647,
+ "learning_rate": 0.0025159014941572785,
+ "loss": 4.826594352722168,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.012433716095983982,
+ "learning_rate": 0.002514916421197171,
+ "loss": 4.797939777374268,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.011882808059453964,
+ "learning_rate": 0.0025139305426224636,
+ "loss": 4.78917932510376,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.011193005368113518,
+ "learning_rate": 0.0025129438592274614,
+ "loss": 4.787964820861816,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.011873407289385796,
+ "learning_rate": 0.002511956371807116,
+ "loss": 4.81754732131958,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.011407418176531792,
+ "learning_rate": 0.002510968081157029,
+ "loss": 4.781639575958252,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.009560304693877697,
+ "learning_rate": 0.002509978988073449,
+ "loss": 4.8163957595825195,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.009969786740839481,
+ "learning_rate": 0.0025089890933532697,
+ "loss": 4.795045852661133,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.010739767923951149,
+ "learning_rate": 0.0025079983977940312,
+ "loss": 4.847584247589111,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.009999158792197704,
+ "learning_rate": 0.0025070069021939194,
+ "loss": 4.834909439086914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.009302482940256596,
+ "learning_rate": 0.002506014607351765,
+ "loss": 4.808745384216309,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.010687744244933128,
+ "learning_rate": 0.0025050215140670403,
+ "loss": 4.789101600646973,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.010930616408586502,
+ "learning_rate": 0.0025040276231398637,
+ "loss": 4.817464351654053,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.013101065531373024,
+ "learning_rate": 0.0025030329353709956,
+ "loss": 4.827780723571777,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.01223307941108942,
+ "learning_rate": 0.0025020374515618373,
+ "loss": 4.8367414474487305,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.013091634958982468,
+ "learning_rate": 0.0025010411725144323,
+ "loss": 4.839454650878906,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.012114843353629112,
+ "learning_rate": 0.0025000440990314643,
+ "loss": 4.832447052001953,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014664588496088982,
+ "learning_rate": 0.0024990462319162588,
+ "loss": 4.740812301635742,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.015566283836960793,
+ "learning_rate": 0.002498047571972777,
+ "loss": 4.815671920776367,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.012440434657037258,
+ "learning_rate": 0.002497048120005623,
+ "loss": 4.769775867462158,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01175499614328146,
+ "learning_rate": 0.0024960478768200362,
+ "loss": 4.8101348876953125,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.012699508108198643,
+ "learning_rate": 0.002495046843221896,
+ "loss": 4.827281951904297,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.012438303790986538,
+ "learning_rate": 0.002494045020017715,
+ "loss": 4.786289691925049,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.01709924265742302,
+ "learning_rate": 0.002493042408014646,
+ "loss": 4.828742027282715,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.022684535011649132,
+ "learning_rate": 0.0024920390080204733,
+ "loss": 4.822258949279785,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.011984334327280521,
+ "learning_rate": 0.00249103482084362,
+ "loss": 4.792006969451904,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.013941102661192417,
+ "learning_rate": 0.00249002984729314,
+ "loss": 4.7926177978515625,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.0145634226500988,
+ "learning_rate": 0.002489024088178724,
+ "loss": 4.819155216217041,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.013391612097620964,
+ "learning_rate": 0.0024880175443106932,
+ "loss": 4.777242660522461,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01764756441116333,
+ "learning_rate": 0.0024870102165000015,
+ "loss": 4.81627893447876,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.015397441573441029,
+ "learning_rate": 0.002486002105558235,
+ "loss": 4.869495391845703,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01225620973855257,
+ "learning_rate": 0.0024849932122976105,
+ "loss": 4.784395217895508,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.01534759160131216,
+ "learning_rate": 0.002483983537530975,
+ "loss": 4.813475608825684,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.015057779848575592,
+ "learning_rate": 0.002482973082071805,
+ "loss": 4.8281569480896,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.01243986189365387,
+ "learning_rate": 0.0024819618467342066,
+ "loss": 4.7866973876953125,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.013326527550816536,
+ "learning_rate": 0.0024809498323329136,
+ "loss": 4.750214576721191,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.013644695281982422,
+ "learning_rate": 0.002479937039683288,
+ "loss": 4.790144920349121,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01296213548630476,
+ "learning_rate": 0.002478923469601319,
+ "loss": 4.812170028686523,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.011848201975226402,
+ "learning_rate": 0.0024779091229036204,
+ "loss": 4.798855781555176,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.012765968218445778,
+ "learning_rate": 0.0024768940004074335,
+ "loss": 4.8272294998168945,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.011700666509568691,
+ "learning_rate": 0.002475878102930625,
+ "loss": 4.762265205383301,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.01018498931080103,
+ "learning_rate": 0.0024748614312916843,
+ "loss": 4.796628952026367,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.010241681709885597,
+ "learning_rate": 0.002473843986309725,
+ "loss": 4.829319000244141,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.009332080371677876,
+ "learning_rate": 0.0024728257688044855,
+ "loss": 4.783773899078369,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.012860726565122604,
+ "learning_rate": 0.002471806779596324,
+ "loss": 4.802700519561768,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.011156697757542133,
+ "learning_rate": 0.0024707870195062225,
+ "loss": 4.808038234710693,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.010329720564186573,
+ "learning_rate": 0.002469766489355783,
+ "loss": 4.761645317077637,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.011093895882368088,
+ "learning_rate": 0.0024687451899672278,
+ "loss": 4.795220851898193,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.010740782134234905,
+ "learning_rate": 0.0024677231221634,
+ "loss": 4.79289436340332,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.012272508814930916,
+ "learning_rate": 0.0024667002867677606,
+ "loss": 4.811873435974121,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.008665879257023335,
+ "learning_rate": 0.0024656766846043905,
+ "loss": 4.744179725646973,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.00933373998850584,
+ "learning_rate": 0.0024646523164979866,
+ "loss": 4.805453777313232,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.010399337857961655,
+ "learning_rate": 0.002463627183273864,
+ "loss": 4.75568962097168,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.013210986740887165,
+ "learning_rate": 0.0024626012857579542,
+ "loss": 4.797974586486816,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.967026233673096,
+ "eval_runtime": 51.07,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.01731858029961586,
+ "learning_rate": 0.0024615746247768043,
+ "loss": 4.748838901519775,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.018902689218521118,
+ "learning_rate": 0.0024605472011575763,
+ "loss": 4.717338562011719,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.01734415628015995,
+ "learning_rate": 0.0024595190157280468,
+ "loss": 4.737737655639648,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.01971321739256382,
+ "learning_rate": 0.002458490069316607,
+ "loss": 4.774558067321777,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.020029844716191292,
+ "learning_rate": 0.002457460362752259,
+ "loss": 4.739579200744629,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.020853109657764435,
+ "learning_rate": 0.0024564298968646203,
+ "loss": 4.791436672210693,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.019159914925694466,
+ "learning_rate": 0.0024553986724839176,
+ "loss": 4.765683650970459,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.0165473110973835,
+ "learning_rate": 0.0024543666904409903,
+ "loss": 4.732603073120117,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.017426826059818268,
+ "learning_rate": 0.002453333951567287,
+ "loss": 4.7360639572143555,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014590734615921974,
+ "learning_rate": 0.002452300456694868,
+ "loss": 4.750031471252441,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.011172222904860973,
+ "learning_rate": 0.0024512662066563996,
+ "loss": 4.740771293640137,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.012294141575694084,
+ "learning_rate": 0.0024502312022851606,
+ "loss": 4.726214408874512,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.01146282535046339,
+ "learning_rate": 0.002449195444415034,
+ "loss": 4.753599166870117,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.011197247542440891,
+ "learning_rate": 0.0024481589338805113,
+ "loss": 4.751479148864746,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.013252820819616318,
+ "learning_rate": 0.0024471216715166904,
+ "loss": 4.743409156799316,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.012366211041808128,
+ "learning_rate": 0.0024460836581592744,
+ "loss": 4.731625556945801,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.013906655833125114,
+ "learning_rate": 0.0024450448946445737,
+ "loss": 4.768153190612793,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.012118754908442497,
+ "learning_rate": 0.002444005381809499,
+ "loss": 4.74198579788208,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.01197550818324089,
+ "learning_rate": 0.002442965120491568,
+ "loss": 4.737289905548096,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.010640372522175312,
+ "learning_rate": 0.002441924111528901,
+ "loss": 4.7472686767578125,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.013553396798670292,
+ "learning_rate": 0.0024408823557602196,
+ "loss": 4.711215972900391,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.01393711194396019,
+ "learning_rate": 0.002439839854024847,
+ "loss": 4.720060348510742,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01589774526655674,
+ "learning_rate": 0.002438796607162709,
+ "loss": 4.695697784423828,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.0097908740863204,
+ "learning_rate": 0.0024377526160143303,
+ "loss": 4.729333877563477,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.011745489202439785,
+ "learning_rate": 0.0024367078814208363,
+ "loss": 4.722957134246826,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.009649662300944328,
+ "learning_rate": 0.0024356624042239498,
+ "loss": 4.727540493011475,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.008429240435361862,
+ "learning_rate": 0.002434616185265993,
+ "loss": 4.705052375793457,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.00984937883913517,
+ "learning_rate": 0.002433569225389887,
+ "loss": 4.721161365509033,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.009237290360033512,
+ "learning_rate": 0.0024325215254391466,
+ "loss": 4.724277496337891,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.008900545537471771,
+ "learning_rate": 0.002431473086257886,
+ "loss": 4.711522102355957,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.010003026574850082,
+ "learning_rate": 0.002430423908690813,
+ "loss": 4.728816032409668,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.011658324860036373,
+ "learning_rate": 0.0024293739935832316,
+ "loss": 4.716179847717285,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.01312185451388359,
+ "learning_rate": 0.0024283233417810393,
+ "loss": 4.7172346115112305,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.014306646771728992,
+ "learning_rate": 0.0024272719541307266,
+ "loss": 4.704054832458496,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.009844634681940079,
+ "learning_rate": 0.0024262198314793783,
+ "loss": 4.72379207611084,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.00903896801173687,
+ "learning_rate": 0.0024251669746746703,
+ "loss": 4.711025238037109,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.010476433672010899,
+ "learning_rate": 0.0024241133845648703,
+ "loss": 4.711664199829102,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.009128391742706299,
+ "learning_rate": 0.0024230590619988375,
+ "loss": 4.72395133972168,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.01228832732886076,
+ "learning_rate": 0.0024220040078260195,
+ "loss": 4.747347831726074,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.011916602961719036,
+ "learning_rate": 0.002420948222896455,
+ "loss": 4.696471691131592,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.012877700850367546,
+ "learning_rate": 0.0024198917080607706,
+ "loss": 4.731119632720947,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.012967764399945736,
+ "learning_rate": 0.0024188344641701814,
+ "loss": 4.688953876495361,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.011993546038866043,
+ "learning_rate": 0.002417776492076489,
+ "loss": 4.695237159729004,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.011188571341335773,
+ "learning_rate": 0.0024167177926320833,
+ "loss": 4.699850082397461,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.010556348599493504,
+ "learning_rate": 0.0024156583666899395,
+ "loss": 4.703063011169434,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.012023619376122952,
+ "learning_rate": 0.0024145982151036176,
+ "loss": 4.6918792724609375,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01266001258045435,
+ "learning_rate": 0.0024135373387272623,
+ "loss": 4.6653289794921875,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.012152725830674171,
+ "learning_rate": 0.0024124757384156024,
+ "loss": 4.682897567749023,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.014056944288313389,
+ "learning_rate": 0.002411413415023951,
+ "loss": 4.696136474609375,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.016826556995511055,
+ "learning_rate": 0.002410350369408202,
+ "loss": 4.701657295227051,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.015048800967633724,
+ "learning_rate": 0.0024092866024248323,
+ "loss": 4.691389083862305,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.012512968853116035,
+ "learning_rate": 0.0024082221149308997,
+ "loss": 4.677886009216309,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01596425101161003,
+ "learning_rate": 0.0024071569077840425,
+ "loss": 4.718337535858154,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01640196517109871,
+ "learning_rate": 0.002406090981842479,
+ "loss": 4.697303295135498,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.012347170151770115,
+ "learning_rate": 0.002405024337965007,
+ "loss": 4.696938991546631,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01213028747588396,
+ "learning_rate": 0.0024039569770110012,
+ "loss": 4.737306594848633,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01314542256295681,
+ "learning_rate": 0.002402888899840416,
+ "loss": 4.69948673248291,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.013764189556241035,
+ "learning_rate": 0.0024018201073137805,
+ "loss": 4.722800254821777,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.011147351004183292,
+ "learning_rate": 0.002400750600292203,
+ "loss": 4.694892883300781,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.015437614172697067,
+ "learning_rate": 0.0023996803796373653,
+ "loss": 4.704752445220947,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.016402097418904305,
+ "learning_rate": 0.0023986094462115255,
+ "loss": 4.67849063873291,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.011600594967603683,
+ "learning_rate": 0.002397537800877514,
+ "loss": 4.689641952514648,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.011864298023283482,
+ "learning_rate": 0.0023964654444987373,
+ "loss": 4.670377731323242,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01409872155636549,
+ "learning_rate": 0.002395392377939173,
+ "loss": 4.696022033691406,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.012106554582715034,
+ "learning_rate": 0.0023943186020633717,
+ "loss": 4.705615043640137,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.012310667894780636,
+ "learning_rate": 0.002393244117736455,
+ "loss": 4.702986717224121,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.013501073233783245,
+ "learning_rate": 0.002392168925824116,
+ "loss": 4.6700215339660645,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.013565126806497574,
+ "learning_rate": 0.0023910930271926166,
+ "loss": 4.699914455413818,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.013421077281236649,
+ "learning_rate": 0.00239001642270879,
+ "loss": 4.713191986083984,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.014408102259039879,
+ "learning_rate": 0.002388939113240035,
+ "loss": 4.655015468597412,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.013502835296094418,
+ "learning_rate": 0.0023878610996543224,
+ "loss": 4.6142258644104,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.014169570058584213,
+ "learning_rate": 0.002386782382820188,
+ "loss": 4.660243034362793,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.885881423950195,
+ "eval_runtime": 52.0315,
+ "eval_samples_per_second": 46.933,
+ "eval_steps_per_second": 1.48,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.014618083834648132,
+ "learning_rate": 0.002385702963606734,
+ "loss": 4.640046119689941,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01745622046291828,
+ "learning_rate": 0.0023846228428836278,
+ "loss": 4.641966819763184,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.014995258301496506,
+ "learning_rate": 0.002383542021521106,
+ "loss": 4.60922908782959,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.013278020545840263,
+ "learning_rate": 0.002382460500389964,
+ "loss": 4.652071952819824,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.014685257337987423,
+ "learning_rate": 0.002381378280361566,
+ "loss": 4.672174453735352,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.012737024575471878,
+ "learning_rate": 0.002380295362307836,
+ "loss": 4.612064361572266,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.012992842122912407,
+ "learning_rate": 0.002379211747101261,
+ "loss": 4.611995697021484,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.010210646316409111,
+ "learning_rate": 0.0023781274356148912,
+ "loss": 4.6055707931518555,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.010265499353408813,
+ "learning_rate": 0.0023770424287223365,
+ "loss": 4.637089729309082,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.010174009948968887,
+ "learning_rate": 0.002375956727297768,
+ "loss": 4.620571136474609,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.009512446820735931,
+ "learning_rate": 0.0023748703322159143,
+ "loss": 4.644092559814453,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.009849252179265022,
+ "learning_rate": 0.0023737832443520655,
+ "loss": 4.625055313110352,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.010493188165128231,
+ "learning_rate": 0.0023726954645820674,
+ "loss": 4.644410610198975,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.022181199863553047,
+ "learning_rate": 0.002371606993782326,
+ "loss": 4.593084335327148,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.011906919069588184,
+ "learning_rate": 0.0023705178328298016,
+ "loss": 4.6452836990356445,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.011874495074152946,
+ "learning_rate": 0.0023694279826020122,
+ "loss": 4.599322319030762,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.00989708211272955,
+ "learning_rate": 0.0023683374439770295,
+ "loss": 4.607342720031738,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.009256849065423012,
+ "learning_rate": 0.0023672462178334817,
+ "loss": 4.621583461761475,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.013187132775783539,
+ "learning_rate": 0.0023661543050505496,
+ "loss": 4.614629745483398,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.013505950570106506,
+ "learning_rate": 0.002365061706507968,
+ "loss": 4.629693984985352,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.012032733298838139,
+ "learning_rate": 0.0023639684230860225,
+ "loss": 4.671209335327148,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.01278337650001049,
+ "learning_rate": 0.002362874455665553,
+ "loss": 4.648874759674072,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.01123690977692604,
+ "learning_rate": 0.0023617798051279496,
+ "loss": 4.611470699310303,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.012564311735332012,
+ "learning_rate": 0.002360684472355151,
+ "loss": 4.659026145935059,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.01206638477742672,
+ "learning_rate": 0.002359588458229649,
+ "loss": 4.642566204071045,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.010300020687282085,
+ "learning_rate": 0.0023584917636344806,
+ "loss": 4.629585266113281,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.0102265365421772,
+ "learning_rate": 0.0023573943894532336,
+ "loss": 4.609250068664551,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.012440572492778301,
+ "learning_rate": 0.002356296336570042,
+ "loss": 4.636714458465576,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.011830241419374943,
+ "learning_rate": 0.0023551976058695887,
+ "loss": 4.59100866317749,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.011482699774205685,
+ "learning_rate": 0.0023540981982370997,
+ "loss": 4.620088577270508,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.012950294651091099,
+ "learning_rate": 0.0023529981145583483,
+ "loss": 4.578657627105713,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.012197710573673248,
+ "learning_rate": 0.002351897355719652,
+ "loss": 4.620813846588135,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.014839230105280876,
+ "learning_rate": 0.002350795922607872,
+ "loss": 4.615731239318848,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016670791432261467,
+ "learning_rate": 0.002349693816110414,
+ "loss": 4.655855178833008,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.014426100067794323,
+ "learning_rate": 0.0023485910371152247,
+ "loss": 4.660849571228027,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.014040560461580753,
+ "learning_rate": 0.0023474875865107936,
+ "loss": 4.634917736053467,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018269391730427742,
+ "learning_rate": 0.0023463834651861496,
+ "loss": 4.611809253692627,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.013855630531907082,
+ "learning_rate": 0.0023452786740308647,
+ "loss": 4.615225791931152,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.011505477130413055,
+ "learning_rate": 0.0023441732139350487,
+ "loss": 4.62592887878418,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.010166114196181297,
+ "learning_rate": 0.002343067085789351,
+ "loss": 4.6219024658203125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.009253021329641342,
+ "learning_rate": 0.0023419602904849596,
+ "loss": 4.658320426940918,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01027526892721653,
+ "learning_rate": 0.002340852828913598,
+ "loss": 4.589777946472168,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.008855022490024567,
+ "learning_rate": 0.002339744701967529,
+ "loss": 4.588204383850098,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.00965463649481535,
+ "learning_rate": 0.002338635910539551,
+ "loss": 4.648177146911621,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.009073720313608646,
+ "learning_rate": 0.0023375264555229956,
+ "loss": 4.585582733154297,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.010105383582413197,
+ "learning_rate": 0.0023364163378117317,
+ "loss": 4.612174987792969,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.015307556837797165,
+ "learning_rate": 0.002335305558300161,
+ "loss": 4.591984748840332,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016716254875063896,
+ "learning_rate": 0.002334194117883219,
+ "loss": 4.659285545349121,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.013281598687171936,
+ "learning_rate": 0.0023330820174563725,
+ "loss": 4.645829200744629,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.014416513964533806,
+ "learning_rate": 0.0023319692579156204,
+ "loss": 4.584893226623535,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.016326071694493294,
+ "learning_rate": 0.002330855840157494,
+ "loss": 4.633223533630371,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.012851824052631855,
+ "learning_rate": 0.002329741765079053,
+ "loss": 4.580986499786377,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.015444506891071796,
+ "learning_rate": 0.0023286270335778876,
+ "loss": 4.615891456604004,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.01391979493200779,
+ "learning_rate": 0.0023275116465521174,
+ "loss": 4.628403186798096,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.012138686142861843,
+ "learning_rate": 0.0023263956049003895,
+ "loss": 4.6259942054748535,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.013741672039031982,
+ "learning_rate": 0.0023252789095218785,
+ "loss": 4.6056742668151855,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01103319600224495,
+ "learning_rate": 0.002324161561316285,
+ "loss": 4.621670246124268,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.010333316400647163,
+ "learning_rate": 0.0023230435611838367,
+ "loss": 4.620484352111816,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.011063606478273869,
+ "learning_rate": 0.0023219249100252864,
+ "loss": 4.64952278137207,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.009984857402741909,
+ "learning_rate": 0.0023208056087419106,
+ "loss": 4.572571754455566,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.009986856952309608,
+ "learning_rate": 0.002319685658235511,
+ "loss": 4.570366382598877,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.010905452072620392,
+ "learning_rate": 0.0023185650594084103,
+ "loss": 4.603614807128906,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.012800061143934727,
+ "learning_rate": 0.002317443813163456,
+ "loss": 4.562705039978027,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.01252608560025692,
+ "learning_rate": 0.002316321920404015,
+ "loss": 4.599016189575195,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.011305592954158783,
+ "learning_rate": 0.002315199382033976,
+ "loss": 4.644617557525635,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.010594413615763187,
+ "learning_rate": 0.002314076198957749,
+ "loss": 4.569808006286621,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.009859138168394566,
+ "learning_rate": 0.002312952372080261,
+ "loss": 4.5890398025512695,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.010982917621731758,
+ "learning_rate": 0.00231182790230696,
+ "loss": 4.606348991394043,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.010252755135297775,
+ "learning_rate": 0.00231070279054381,
+ "loss": 4.58942174911499,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01110257301479578,
+ "learning_rate": 0.0023095770376972947,
+ "loss": 4.599823474884033,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01003815233707428,
+ "learning_rate": 0.0023084506446744103,
+ "loss": 4.569343090057373,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.010517258197069168,
+ "learning_rate": 0.0023073236123826745,
+ "loss": 4.646301746368408,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.804611682891846,
+ "eval_runtime": 51.6279,
+ "eval_samples_per_second": 47.3,
+ "eval_steps_per_second": 1.491,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.011439066380262375,
+ "learning_rate": 0.0023061959417301145,
+ "loss": 4.563335418701172,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.01073157787322998,
+ "learning_rate": 0.002305067633625275,
+ "loss": 4.540966987609863,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.011391228064894676,
+ "learning_rate": 0.0023039386889772136,
+ "loss": 4.555583953857422,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.012814595364034176,
+ "learning_rate": 0.0023028091086955,
+ "loss": 4.517425537109375,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.012994988821446896,
+ "learning_rate": 0.002301678893690218,
+ "loss": 4.538477897644043,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.01460991520434618,
+ "learning_rate": 0.00230054804487196,
+ "loss": 4.571145057678223,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.013036075979471207,
+ "learning_rate": 0.0022994165631518306,
+ "loss": 4.519233703613281,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.011917057447135448,
+ "learning_rate": 0.0022982844494414444,
+ "loss": 4.5071587562561035,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.013928866013884544,
+ "learning_rate": 0.0022971517046529257,
+ "loss": 4.531422138214111,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01407038513571024,
+ "learning_rate": 0.0022960183296989053,
+ "loss": 4.511205196380615,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.019396567717194557,
+ "learning_rate": 0.0022948843254925245,
+ "loss": 4.530007362365723,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.014425835572183132,
+ "learning_rate": 0.002293749692947428,
+ "loss": 4.530138969421387,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.01600247249007225,
+ "learning_rate": 0.00229261443297777,
+ "loss": 4.5643391609191895,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.018238037824630737,
+ "learning_rate": 0.0022914785464982095,
+ "loss": 4.540779113769531,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.015427535399794579,
+ "learning_rate": 0.0022903420344239085,
+ "loss": 4.506287097930908,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.016816625371575356,
+ "learning_rate": 0.0022892048976705347,
+ "loss": 4.537919521331787,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.01999717950820923,
+ "learning_rate": 0.002288067137154259,
+ "loss": 4.558426380157471,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.020884918048977852,
+ "learning_rate": 0.002286928753791753,
+ "loss": 4.592915058135986,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.01620141603052616,
+ "learning_rate": 0.002285789748500194,
+ "loss": 4.553356170654297,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.018495604395866394,
+ "learning_rate": 0.0022846501221972567,
+ "loss": 4.573139190673828,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.0156320258975029,
+ "learning_rate": 0.0022835098758011174,
+ "loss": 4.585921764373779,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.014515317976474762,
+ "learning_rate": 0.002282369010230452,
+ "loss": 4.5532002449035645,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.014826556667685509,
+ "learning_rate": 0.0022812275264044355,
+ "loss": 4.560998916625977,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.013240831904113293,
+ "learning_rate": 0.00228008542524274,
+ "loss": 4.531787872314453,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.013814719393849373,
+ "learning_rate": 0.0022789427076655372,
+ "loss": 4.611872673034668,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.012246561236679554,
+ "learning_rate": 0.0022777993745934934,
+ "loss": 4.5565385818481445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01206218171864748,
+ "learning_rate": 0.002276655426947771,
+ "loss": 4.570852279663086,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01007329672574997,
+ "learning_rate": 0.0022755108656500275,
+ "loss": 4.553400039672852,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.010361720807850361,
+ "learning_rate": 0.0022743656916224173,
+ "loss": 4.527509689331055,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.009456256404519081,
+ "learning_rate": 0.0022732199057875846,
+ "loss": 4.520236015319824,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.009310049936175346,
+ "learning_rate": 0.002272073509068669,
+ "loss": 4.518343925476074,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.00927449855953455,
+ "learning_rate": 0.0022709265023893015,
+ "loss": 4.559690475463867,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.008658594451844692,
+ "learning_rate": 0.002269778886673605,
+ "loss": 4.513876914978027,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.008928220719099045,
+ "learning_rate": 0.0022686306628461918,
+ "loss": 4.545050621032715,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.008046358823776245,
+ "learning_rate": 0.002267481831832166,
+ "loss": 4.54866361618042,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.008318924345076084,
+ "learning_rate": 0.00226633239455712,
+ "loss": 4.546968936920166,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.008808869868516922,
+ "learning_rate": 0.0022651823519471343,
+ "loss": 4.50877571105957,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.008767383173108101,
+ "learning_rate": 0.0022640317049287774,
+ "loss": 4.57666540145874,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.009351354092359543,
+ "learning_rate": 0.0022628804544291047,
+ "loss": 4.511702537536621,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.010137057863175869,
+ "learning_rate": 0.002261728601375659,
+ "loss": 4.539676189422607,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.00882046204060316,
+ "learning_rate": 0.002260576146696466,
+ "loss": 4.533196449279785,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.009354782290756702,
+ "learning_rate": 0.0022594230913200388,
+ "loss": 4.515488624572754,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.009241201914846897,
+ "learning_rate": 0.002258269436175372,
+ "loss": 4.542706489562988,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.009252829477190971,
+ "learning_rate": 0.002257115182191947,
+ "loss": 4.592660903930664,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.011243131943047047,
+ "learning_rate": 0.0022559603302997225,
+ "loss": 4.527804851531982,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.011032823473215103,
+ "learning_rate": 0.0022548048814291437,
+ "loss": 4.540122032165527,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.010495275259017944,
+ "learning_rate": 0.0022536488365111347,
+ "loss": 4.532835006713867,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.009992090053856373,
+ "learning_rate": 0.0022524921964770994,
+ "loss": 4.521217346191406,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01141198631376028,
+ "learning_rate": 0.0022513349622589225,
+ "loss": 4.519709587097168,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.013087889179587364,
+ "learning_rate": 0.002250177134788966,
+ "loss": 4.547289848327637,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.01328477542847395,
+ "learning_rate": 0.0022490187150000706,
+ "loss": 4.527865409851074,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01033794041723013,
+ "learning_rate": 0.002247859703825555,
+ "loss": 4.54740047454834,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.010522300377488136,
+ "learning_rate": 0.002246700102199213,
+ "loss": 4.538840293884277,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.012151090428233147,
+ "learning_rate": 0.0022455399110553145,
+ "loss": 4.536865234375,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.010553102008998394,
+ "learning_rate": 0.0022443791313286054,
+ "loss": 4.5348334312438965,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.011433709412813187,
+ "learning_rate": 0.0022432177639543035,
+ "loss": 4.510526657104492,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01189691387116909,
+ "learning_rate": 0.002242055809868103,
+ "loss": 4.530807018280029,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.010031254962086678,
+ "learning_rate": 0.0022408932700061683,
+ "loss": 4.496850967407227,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.010129331611096859,
+ "learning_rate": 0.002239730145305138,
+ "loss": 4.551287651062012,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.010430586524307728,
+ "learning_rate": 0.0022385664367021185,
+ "loss": 4.542672157287598,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.011774476617574692,
+ "learning_rate": 0.002237402145134691,
+ "loss": 4.537034034729004,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.01184659544378519,
+ "learning_rate": 0.002236237271540903,
+ "loss": 4.521193504333496,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.011327247135341167,
+ "learning_rate": 0.0022350718168592724,
+ "loss": 4.49426794052124,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.010767918080091476,
+ "learning_rate": 0.0022339057820287855,
+ "loss": 4.531225204467773,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.011195424944162369,
+ "learning_rate": 0.002232739167988894,
+ "loss": 4.492671012878418,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.010394908487796783,
+ "learning_rate": 0.0022315719756795197,
+ "loss": 4.523616790771484,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.010277998633682728,
+ "learning_rate": 0.002230404206041047,
+ "loss": 4.487832069396973,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.010955783538520336,
+ "learning_rate": 0.0022292358600143275,
+ "loss": 4.526865005493164,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.010611076839268208,
+ "learning_rate": 0.002228066938540676,
+ "loss": 4.498682022094727,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.010008828714489937,
+ "learning_rate": 0.0022268974425618724,
+ "loss": 4.524019241333008,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.008973540738224983,
+ "learning_rate": 0.002225727373020157,
+ "loss": 4.512996673583984,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.011754963546991348,
+ "learning_rate": 0.002224556730858235,
+ "loss": 4.496376991271973,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.744190216064453,
+ "eval_runtime": 51.6572,
+ "eval_samples_per_second": 47.273,
+ "eval_steps_per_second": 1.491,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.012094628997147083,
+ "learning_rate": 0.0022233855170192703,
+ "loss": 4.443447113037109,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.014347990974783897,
+ "learning_rate": 0.00222221373244689,
+ "loss": 4.485690116882324,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.021286984905600548,
+ "learning_rate": 0.0022210413780851786,
+ "loss": 4.487505912780762,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.020273910835385323,
+ "learning_rate": 0.0022198684548786827,
+ "loss": 4.473119735717773,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.017642971128225327,
+ "learning_rate": 0.002218694963772403,
+ "loss": 4.440615653991699,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.02075147069990635,
+ "learning_rate": 0.0022175209057118017,
+ "loss": 4.527026176452637,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016479508951306343,
+ "learning_rate": 0.002216346281642796,
+ "loss": 4.508321762084961,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.017978593707084656,
+ "learning_rate": 0.002215171092511759,
+ "loss": 4.500985145568848,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.01739897020161152,
+ "learning_rate": 0.002213995339265519,
+ "loss": 4.474225044250488,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.015556776896119118,
+ "learning_rate": 0.0022128190228513597,
+ "loss": 4.444849014282227,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.016166578978300095,
+ "learning_rate": 0.002211642144217018,
+ "loss": 4.5026726722717285,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.01534261740744114,
+ "learning_rate": 0.0022104647043106832,
+ "loss": 4.484527587890625,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.013517106883227825,
+ "learning_rate": 0.002209286704080998,
+ "loss": 4.453907012939453,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.012562787160277367,
+ "learning_rate": 0.0022081081444770556,
+ "loss": 4.471528053283691,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.009947682730853558,
+ "learning_rate": 0.0022069290264483997,
+ "loss": 4.504153728485107,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01169571466743946,
+ "learning_rate": 0.002205749350945026,
+ "loss": 4.447345733642578,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.011394352652132511,
+ "learning_rate": 0.0022045691189173755,
+ "loss": 4.447491645812988,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.011168593540787697,
+ "learning_rate": 0.0022033883313163417,
+ "loss": 4.441639423370361,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.009503341279923916,
+ "learning_rate": 0.002202206989093263,
+ "loss": 4.461601257324219,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.0112199941650033,
+ "learning_rate": 0.002201025093199926,
+ "loss": 4.444095134735107,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.010505333542823792,
+ "learning_rate": 0.002199842644588562,
+ "loss": 4.460987091064453,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.009494813159108162,
+ "learning_rate": 0.0021986596442118496,
+ "loss": 4.460158348083496,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.01063452661037445,
+ "learning_rate": 0.00219747609302291,
+ "loss": 4.489625453948975,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.012194888666272163,
+ "learning_rate": 0.0021962919919753087,
+ "loss": 4.441605567932129,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.010613647289574146,
+ "learning_rate": 0.0021951073420230557,
+ "loss": 4.475821018218994,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.009643700905144215,
+ "learning_rate": 0.002193922144120601,
+ "loss": 4.467228889465332,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.00912613607943058,
+ "learning_rate": 0.0021927363992228373,
+ "loss": 4.455435752868652,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.009371735155582428,
+ "learning_rate": 0.0021915501082850988,
+ "loss": 4.479915142059326,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.008617875166237354,
+ "learning_rate": 0.002190363272263158,
+ "loss": 4.415642738342285,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.009571383707225323,
+ "learning_rate": 0.002189175892113227,
+ "loss": 4.439854621887207,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.00782635249197483,
+ "learning_rate": 0.002187987968791957,
+ "loss": 4.484311103820801,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.008408057503402233,
+ "learning_rate": 0.0021867995032564358,
+ "loss": 4.46577262878418,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.009641910903155804,
+ "learning_rate": 0.00218561049646419,
+ "loss": 4.437408447265625,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.008498182520270348,
+ "learning_rate": 0.0021844209493731784,
+ "loss": 4.442599296569824,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.009289798326790333,
+ "learning_rate": 0.0021832308629418,
+ "loss": 4.517641067504883,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.009360183961689472,
+ "learning_rate": 0.0021820402381288848,
+ "loss": 4.474143981933594,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.008764738216996193,
+ "learning_rate": 0.0021808490758936983,
+ "loss": 4.4376702308654785,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.009673701599240303,
+ "learning_rate": 0.0021796573771959375,
+ "loss": 4.435220718383789,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.010852573439478874,
+ "learning_rate": 0.0021784651429957334,
+ "loss": 4.465187072753906,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.011292314156889915,
+ "learning_rate": 0.0021772723742536464,
+ "loss": 4.458176136016846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01071334257721901,
+ "learning_rate": 0.0021760790719306702,
+ "loss": 4.425498962402344,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.01505239401012659,
+ "learning_rate": 0.0021748852369882265,
+ "loss": 4.463129043579102,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.011774689890444279,
+ "learning_rate": 0.0021736908703881663,
+ "loss": 4.462577819824219,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.011241118423640728,
+ "learning_rate": 0.002172495973092769,
+ "loss": 4.452583312988281,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.01072910986840725,
+ "learning_rate": 0.002171300546064742,
+ "loss": 4.466248035430908,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01021308358758688,
+ "learning_rate": 0.002170104590267219,
+ "loss": 4.485857963562012,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.012630444951355457,
+ "learning_rate": 0.00216890810666376,
+ "loss": 4.442112922668457,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.013602569699287415,
+ "learning_rate": 0.0021677110962183515,
+ "loss": 4.45171594619751,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.013808641582727432,
+ "learning_rate": 0.0021665135598954014,
+ "loss": 4.468818187713623,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.011510699987411499,
+ "learning_rate": 0.002165315498659744,
+ "loss": 4.439584255218506,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.010502424091100693,
+ "learning_rate": 0.0021641169134766353,
+ "loss": 4.478377342224121,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.009953407570719719,
+ "learning_rate": 0.0021629178053117542,
+ "loss": 4.474404335021973,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.011158930137753487,
+ "learning_rate": 0.0021617181751312,
+ "loss": 4.4549560546875,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.010431940667331219,
+ "learning_rate": 0.002160518023901493,
+ "loss": 4.444911003112793,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.009100335650146008,
+ "learning_rate": 0.0021593173525895734,
+ "loss": 4.428071975708008,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.009362847544252872,
+ "learning_rate": 0.0021581161621628013,
+ "loss": 4.462827682495117,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.009893129579722881,
+ "learning_rate": 0.0021569144535889535,
+ "loss": 4.426912307739258,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.010524468496441841,
+ "learning_rate": 0.0021557122278362244,
+ "loss": 4.435606002807617,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.010621445253491402,
+ "learning_rate": 0.002154509485873226,
+ "loss": 4.454901695251465,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.010836896486580372,
+ "learning_rate": 0.0021533062286689855,
+ "loss": 4.4441728591918945,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.011403030715882778,
+ "learning_rate": 0.0021521024571929457,
+ "loss": 4.483884811401367,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.012522426433861256,
+ "learning_rate": 0.0021508981724149643,
+ "loss": 4.464006423950195,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.014799179509282112,
+ "learning_rate": 0.0021496933753053105,
+ "loss": 4.460402011871338,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.012453878298401833,
+ "learning_rate": 0.0021484880668346685,
+ "loss": 4.456899642944336,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.012270379811525345,
+ "learning_rate": 0.002147282247974133,
+ "loss": 4.477325439453125,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01094560045748949,
+ "learning_rate": 0.0021460759196952105,
+ "loss": 4.422362327575684,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.009995176456868649,
+ "learning_rate": 0.0021448690829698183,
+ "loss": 4.405127048492432,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.014935536310076714,
+ "learning_rate": 0.002143661738770283,
+ "loss": 4.445840358734131,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.011492030695080757,
+ "learning_rate": 0.0021424538880693393,
+ "loss": 4.459555625915527,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.008793608285486698,
+ "learning_rate": 0.002141245531840132,
+ "loss": 4.436631202697754,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.008894025348126888,
+ "learning_rate": 0.0021400366710562103,
+ "loss": 4.473924160003662,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.010324863716959953,
+ "learning_rate": 0.0021388273066915327,
+ "loss": 4.475680351257324,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.688324451446533,
+ "eval_runtime": 54.4515,
+ "eval_samples_per_second": 44.847,
+ "eval_steps_per_second": 1.414,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.009235628880560398,
+ "learning_rate": 0.0021376174397204615,
+ "loss": 4.370064735412598,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.008818200789391994,
+ "learning_rate": 0.0021364070711177643,
+ "loss": 4.371678352355957,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.010333757847547531,
+ "learning_rate": 0.0021351962018586147,
+ "loss": 4.3896379470825195,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.011938809417188168,
+ "learning_rate": 0.002133984832918587,
+ "loss": 4.361343860626221,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.010634614154696465,
+ "learning_rate": 0.00213277296527366,
+ "loss": 4.362449645996094,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.008545687422156334,
+ "learning_rate": 0.0021315605999002126,
+ "loss": 4.3685503005981445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.009159795008599758,
+ "learning_rate": 0.002130347737775027,
+ "loss": 4.379969596862793,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.010464153252542019,
+ "learning_rate": 0.0021291343798752835,
+ "loss": 4.376113414764404,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.011010527610778809,
+ "learning_rate": 0.0021279205271785622,
+ "loss": 4.390711784362793,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.010350867174565792,
+ "learning_rate": 0.002126706180662843,
+ "loss": 4.399683952331543,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.009708237834274769,
+ "learning_rate": 0.002125491341306503,
+ "loss": 4.363435745239258,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.010087115690112114,
+ "learning_rate": 0.002124276010088316,
+ "loss": 4.366623878479004,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.009712349623441696,
+ "learning_rate": 0.002123060187987453,
+ "loss": 4.360814571380615,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.010174986906349659,
+ "learning_rate": 0.0021218438759834785,
+ "loss": 4.347475051879883,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.010195785202085972,
+ "learning_rate": 0.0021206270750563543,
+ "loss": 4.399251461029053,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.010786959901452065,
+ "learning_rate": 0.0021194097861864354,
+ "loss": 4.397933006286621,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.011476189829409122,
+ "learning_rate": 0.002118192010354468,
+ "loss": 4.3748369216918945,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.011571013368666172,
+ "learning_rate": 0.0021169737485415927,
+ "loss": 4.388349533081055,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.011140935122966766,
+ "learning_rate": 0.002115755001729341,
+ "loss": 4.398018836975098,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.013216734863817692,
+ "learning_rate": 0.0021145357708996365,
+ "loss": 4.406947135925293,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.013385126367211342,
+ "learning_rate": 0.0021133160570347893,
+ "loss": 4.41457462310791,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.011784231290221214,
+ "learning_rate": 0.002112095861117502,
+ "loss": 4.403835296630859,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.009813498705625534,
+ "learning_rate": 0.0021108751841308645,
+ "loss": 4.3865647315979,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.010297936387360096,
+ "learning_rate": 0.0021096540270583533,
+ "loss": 4.395583152770996,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.011965147219598293,
+ "learning_rate": 0.002108432390883834,
+ "loss": 4.385443687438965,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.012212193571031094,
+ "learning_rate": 0.0021072102765915556,
+ "loss": 4.422321319580078,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.009840975515544415,
+ "learning_rate": 0.0021059876851661544,
+ "loss": 4.3568902015686035,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.013039806857705116,
+ "learning_rate": 0.002104764617592649,
+ "loss": 4.314867973327637,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.015091902576386929,
+ "learning_rate": 0.0021035410748564444,
+ "loss": 4.392338752746582,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.015448705293238163,
+ "learning_rate": 0.0021023170579433257,
+ "loss": 4.379609107971191,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.018393104895949364,
+ "learning_rate": 0.0021010925678394622,
+ "loss": 4.418291091918945,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.01866372860968113,
+ "learning_rate": 0.002099867605531402,
+ "loss": 4.4006452560424805,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.01819106750190258,
+ "learning_rate": 0.002098642172006077,
+ "loss": 4.433975696563721,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.019774602726101875,
+ "learning_rate": 0.0020974162682507952,
+ "loss": 4.417821884155273,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.016885802149772644,
+ "learning_rate": 0.0020961898952532457,
+ "loss": 4.397073268890381,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.015750881284475327,
+ "learning_rate": 0.0020949630540014956,
+ "loss": 4.4356608390808105,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.015696514397859573,
+ "learning_rate": 0.002093735745483988,
+ "loss": 4.422449588775635,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.017195021733641624,
+ "learning_rate": 0.0020925079706895435,
+ "loss": 4.443209648132324,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.015864048153162003,
+ "learning_rate": 0.0020912797306073577,
+ "loss": 4.4340691566467285,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.009857800789177418,
+ "learning_rate": 0.0020900510262270022,
+ "loss": 4.37935209274292,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.013739476911723614,
+ "learning_rate": 0.0020888218585384216,
+ "loss": 4.412148475646973,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.011667690239846706,
+ "learning_rate": 0.0020875922285319344,
+ "loss": 4.391700744628906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.012323460541665554,
+ "learning_rate": 0.0020863621371982302,
+ "loss": 4.423880577087402,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.010353626683354378,
+ "learning_rate": 0.0020851315855283724,
+ "loss": 4.427382469177246,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.0101751908659935,
+ "learning_rate": 0.002083900574513794,
+ "loss": 4.415317535400391,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.01126549206674099,
+ "learning_rate": 0.002082669105146298,
+ "loss": 4.402498245239258,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.011096738278865814,
+ "learning_rate": 0.0020814371784180575,
+ "loss": 4.433567047119141,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.010944445617496967,
+ "learning_rate": 0.0020802047953216137,
+ "loss": 4.406990051269531,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.008965318091213703,
+ "learning_rate": 0.0020789719568498747,
+ "loss": 4.3721771240234375,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.00828113779425621,
+ "learning_rate": 0.002077738663996117,
+ "loss": 4.389828205108643,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.008599454537034035,
+ "learning_rate": 0.002076504917753982,
+ "loss": 4.411396026611328,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.007713889703154564,
+ "learning_rate": 0.0020752707191174762,
+ "loss": 4.355343341827393,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.00912355538457632,
+ "learning_rate": 0.0020740360690809723,
+ "loss": 4.423413276672363,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.008269775658845901,
+ "learning_rate": 0.0020728009686392045,
+ "loss": 4.390920639038086,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.007461601868271828,
+ "learning_rate": 0.002071565418787272,
+ "loss": 4.365501403808594,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.007951333187520504,
+ "learning_rate": 0.0020703294205206336,
+ "loss": 4.378884315490723,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.007816503755748272,
+ "learning_rate": 0.002069092974835112,
+ "loss": 4.4241743087768555,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.00833756010979414,
+ "learning_rate": 0.0020678560827268886,
+ "loss": 4.391102313995361,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.008241684176027775,
+ "learning_rate": 0.0020666187451925055,
+ "loss": 4.374970436096191,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.008245996199548244,
+ "learning_rate": 0.002065380963228862,
+ "loss": 4.380222320556641,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.007378706708550453,
+ "learning_rate": 0.0020641427378332183,
+ "loss": 4.391451358795166,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.007594958413392305,
+ "learning_rate": 0.002062904070003189,
+ "loss": 4.372774124145508,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.0076752156019210815,
+ "learning_rate": 0.002061664960736747,
+ "loss": 4.384668827056885,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.008727089501917362,
+ "learning_rate": 0.00206042541103222,
+ "loss": 4.363401412963867,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.007616396062076092,
+ "learning_rate": 0.002059185421888291,
+ "loss": 4.413188934326172,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.008922954089939594,
+ "learning_rate": 0.0020579449943039967,
+ "loss": 4.383417129516602,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.009063739329576492,
+ "learning_rate": 0.002056704129278727,
+ "loss": 4.385283470153809,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.008912574499845505,
+ "learning_rate": 0.002055462827812225,
+ "loss": 4.378478050231934,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.008317525498569012,
+ "learning_rate": 0.0020542210909045844,
+ "loss": 4.367859840393066,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.009347227402031422,
+ "learning_rate": 0.0020529789195562506,
+ "loss": 4.38238000869751,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.011030065827071667,
+ "learning_rate": 0.0020517363147680183,
+ "loss": 4.372624397277832,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.011519700288772583,
+ "learning_rate": 0.002050493277541032,
+ "loss": 4.3906121253967285,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.641542434692383,
+ "eval_runtime": 51.6342,
+ "eval_samples_per_second": 47.294,
+ "eval_steps_per_second": 1.491,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.009694516658782959,
+ "learning_rate": 0.0020492498088767836,
+ "loss": 4.32808256149292,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.0097952950745821,
+ "learning_rate": 0.002048005909777114,
+ "loss": 4.280096054077148,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.010393359698355198,
+ "learning_rate": 0.0020467615812442106,
+ "loss": 4.317824363708496,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.00977372657507658,
+ "learning_rate": 0.0020455168242806054,
+ "loss": 4.3445329666137695,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.008568719029426575,
+ "learning_rate": 0.0020442716398891775,
+ "loss": 4.318026542663574,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.009732510894536972,
+ "learning_rate": 0.0020430260290731487,
+ "loss": 4.330609321594238,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.010429969057440758,
+ "learning_rate": 0.0020417799928360864,
+ "loss": 4.316794395446777,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.009292383678257465,
+ "learning_rate": 0.002040533532181898,
+ "loss": 4.317420959472656,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.011222455650568008,
+ "learning_rate": 0.0020392866481148356,
+ "loss": 4.292610168457031,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.013591725379228592,
+ "learning_rate": 0.0020380393416394894,
+ "loss": 4.325689315795898,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.013572026044130325,
+ "learning_rate": 0.002036791613760794,
+ "loss": 4.334960460662842,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.012986521236598492,
+ "learning_rate": 0.0020355434654840197,
+ "loss": 4.33829402923584,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.013321424834430218,
+ "learning_rate": 0.0020342948978147777,
+ "loss": 4.2926926612854,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.012307215481996536,
+ "learning_rate": 0.0020330459117590153,
+ "loss": 4.32757568359375,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.012909360229969025,
+ "learning_rate": 0.0020317965083230197,
+ "loss": 4.362889289855957,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017746610566973686,
+ "learning_rate": 0.0020305466885134115,
+ "loss": 4.3426313400268555,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.01696564070880413,
+ "learning_rate": 0.002029296453337148,
+ "loss": 4.3298516273498535,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.013348255306482315,
+ "learning_rate": 0.002028045803801522,
+ "loss": 4.360220432281494,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.013225145637989044,
+ "learning_rate": 0.002026794740914158,
+ "loss": 4.346250057220459,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.01569020375609398,
+ "learning_rate": 0.002025543265683016,
+ "loss": 4.308025360107422,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.018612321466207504,
+ "learning_rate": 0.002024291379116385,
+ "loss": 4.330190181732178,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.01619943603873253,
+ "learning_rate": 0.00202303908222289,
+ "loss": 4.339552402496338,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.015385213308036327,
+ "learning_rate": 0.0020217863760114826,
+ "loss": 4.3412184715271,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.018517112359404564,
+ "learning_rate": 0.002020533261491446,
+ "loss": 4.3289313316345215,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.014868509024381638,
+ "learning_rate": 0.0020192797396723903,
+ "loss": 4.359846115112305,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.014240248128771782,
+ "learning_rate": 0.0020180258115642578,
+ "loss": 4.369707107543945,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.018113158643245697,
+ "learning_rate": 0.0020167714781773143,
+ "loss": 4.362668514251709,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.01802460104227066,
+ "learning_rate": 0.002015516740522154,
+ "loss": 4.344742298126221,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.01414549071341753,
+ "learning_rate": 0.0020142615996096964,
+ "loss": 4.358080863952637,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.012157700955867767,
+ "learning_rate": 0.002013006056451186,
+ "loss": 4.347763538360596,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.011268534697592258,
+ "learning_rate": 0.0020117501120581905,
+ "loss": 4.335859298706055,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.010395628400146961,
+ "learning_rate": 0.002010493767442602,
+ "loss": 4.341609001159668,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.012147950939834118,
+ "learning_rate": 0.002009237023616635,
+ "loss": 4.328502655029297,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.00945813488215208,
+ "learning_rate": 0.0020079798815928252,
+ "loss": 4.311186790466309,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.009489500895142555,
+ "learning_rate": 0.0020067223423840294,
+ "loss": 4.371470928192139,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.009360029362142086,
+ "learning_rate": 0.0020054644070034234,
+ "loss": 4.33941125869751,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.010032779537141323,
+ "learning_rate": 0.0020042060764645037,
+ "loss": 4.365797519683838,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.009367046877741814,
+ "learning_rate": 0.002002947351781084,
+ "loss": 4.29443359375,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.008369733579456806,
+ "learning_rate": 0.0020016882339672967,
+ "loss": 4.294233322143555,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.010394785553216934,
+ "learning_rate": 0.0020004287240375893,
+ "loss": 4.352355003356934,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01068816613405943,
+ "learning_rate": 0.001999168823006728,
+ "loss": 4.342516899108887,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.009975973516702652,
+ "learning_rate": 0.0019979085318897905,
+ "loss": 4.346763610839844,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.009618151932954788,
+ "learning_rate": 0.001996647851702171,
+ "loss": 4.329115867614746,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.008399694226682186,
+ "learning_rate": 0.001995386783459577,
+ "loss": 4.28736686706543,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.008235268294811249,
+ "learning_rate": 0.001994125328178029,
+ "loss": 4.339872360229492,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.00836547277867794,
+ "learning_rate": 0.0019928634868738577,
+ "loss": 4.328156471252441,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.009619626216590405,
+ "learning_rate": 0.001991601260563706,
+ "loss": 4.337394714355469,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.009814038872718811,
+ "learning_rate": 0.0019903386502645275,
+ "loss": 4.325926780700684,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.009588954970240593,
+ "learning_rate": 0.001989075656993584,
+ "loss": 4.286477088928223,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.010471943765878677,
+ "learning_rate": 0.001987812281768447,
+ "loss": 4.322466850280762,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.010239035822451115,
+ "learning_rate": 0.001986548525606994,
+ "loss": 4.3487653732299805,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.008935568854212761,
+ "learning_rate": 0.001985284389527411,
+ "loss": 4.303601264953613,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.008624759502708912,
+ "learning_rate": 0.00198401987454819,
+ "loss": 4.34036922454834,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.009296547621488571,
+ "learning_rate": 0.0019827549816881277,
+ "loss": 4.273357391357422,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.009071138687431812,
+ "learning_rate": 0.0019814897119663248,
+ "loss": 4.321051597595215,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.008341389708220959,
+ "learning_rate": 0.0019802240664021876,
+ "loss": 4.343242645263672,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.007881715893745422,
+ "learning_rate": 0.0019789580460154227,
+ "loss": 4.32067346572876,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.008750553242862225,
+ "learning_rate": 0.0019776916518260404,
+ "loss": 4.330263614654541,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.008591460064053535,
+ "learning_rate": 0.001976424884854351,
+ "loss": 4.311957359313965,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.008084543980658054,
+ "learning_rate": 0.001975157746120966,
+ "loss": 4.331319808959961,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.008036594837903976,
+ "learning_rate": 0.001973890236646797,
+ "loss": 4.382656097412109,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.00805772002786398,
+ "learning_rate": 0.0019726223574530527,
+ "loss": 4.297247886657715,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.009019420482218266,
+ "learning_rate": 0.0019713541095612405,
+ "loss": 4.326526165008545,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.009953883476555347,
+ "learning_rate": 0.0019700854939931653,
+ "loss": 4.327356338500977,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.009340492077171803,
+ "learning_rate": 0.0019688165117709273,
+ "loss": 4.343993186950684,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.010395721532404423,
+ "learning_rate": 0.0019675471639169226,
+ "loss": 4.284934997558594,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.009946230798959732,
+ "learning_rate": 0.0019662774514538415,
+ "loss": 4.288936614990234,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.010981100611388683,
+ "learning_rate": 0.0019650073754046695,
+ "loss": 4.339935779571533,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.009688783437013626,
+ "learning_rate": 0.0019637369367926827,
+ "loss": 4.381425857543945,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.009442255832254887,
+ "learning_rate": 0.001962466136641452,
+ "loss": 4.304185390472412,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.009371291846036911,
+ "learning_rate": 0.0019611949759748364,
+ "loss": 4.306120872497559,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.009544814005494118,
+ "learning_rate": 0.001959923455816988,
+ "loss": 4.326251983642578,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.602745532989502,
+ "eval_runtime": 53.5041,
+ "eval_samples_per_second": 45.641,
+ "eval_steps_per_second": 1.439,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.01010364294052124,
+ "learning_rate": 0.0019586515771923474,
+ "loss": 4.24491024017334,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.012722729705274105,
+ "learning_rate": 0.0019573793411256446,
+ "loss": 4.257739543914795,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.01307204831391573,
+ "learning_rate": 0.0019561067486418967,
+ "loss": 4.291684150695801,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.011215966194868088,
+ "learning_rate": 0.0019548338007664098,
+ "loss": 4.237947463989258,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.012943019159138203,
+ "learning_rate": 0.0019535604985247727,
+ "loss": 4.242859363555908,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.01337282732129097,
+ "learning_rate": 0.0019522868429428636,
+ "loss": 4.238934516906738,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.013697991147637367,
+ "learning_rate": 0.0019510128350468438,
+ "loss": 4.285760402679443,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.012182159349322319,
+ "learning_rate": 0.001949738475863158,
+ "loss": 4.290044784545898,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.012620339170098305,
+ "learning_rate": 0.0019484637664185346,
+ "loss": 4.267187118530273,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.011284958571195602,
+ "learning_rate": 0.0019471887077399837,
+ "loss": 4.26662540435791,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.01161450706422329,
+ "learning_rate": 0.001945913300854797,
+ "loss": 4.283192157745361,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.010022968985140324,
+ "learning_rate": 0.001944637546790548,
+ "loss": 4.295843124389648,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.010341878980398178,
+ "learning_rate": 0.001943361446575087,
+ "loss": 4.295636177062988,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.011549185030162334,
+ "learning_rate": 0.0019420850012365457,
+ "loss": 4.288768768310547,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.012518810108304024,
+ "learning_rate": 0.0019408082118033331,
+ "loss": 4.229338645935059,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.012035328894853592,
+ "learning_rate": 0.0019395310793041353,
+ "loss": 4.258639335632324,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.011193782091140747,
+ "learning_rate": 0.0019382536047679146,
+ "loss": 4.239264488220215,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.009930136613547802,
+ "learning_rate": 0.0019369757892239096,
+ "loss": 4.252286434173584,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.00953229982405901,
+ "learning_rate": 0.0019356976337016327,
+ "loss": 4.293435096740723,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.010618340224027634,
+ "learning_rate": 0.0019344191392308716,
+ "loss": 4.235588550567627,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.008806651458144188,
+ "learning_rate": 0.0019331403068416856,
+ "loss": 4.26716423034668,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.008934796787798405,
+ "learning_rate": 0.0019318611375644065,
+ "loss": 4.258462905883789,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.009283283725380898,
+ "learning_rate": 0.0019305816324296386,
+ "loss": 4.257868766784668,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.009224812500178814,
+ "learning_rate": 0.0019293017924682555,
+ "loss": 4.258486747741699,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.008957969956099987,
+ "learning_rate": 0.0019280216187114016,
+ "loss": 4.309563636779785,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.010284051299095154,
+ "learning_rate": 0.0019267411121904894,
+ "loss": 4.275271415710449,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.009167762473225594,
+ "learning_rate": 0.0019254602739372001,
+ "loss": 4.238452911376953,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.009411636739969254,
+ "learning_rate": 0.0019241791049834816,
+ "loss": 4.2735595703125,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.009044171310961246,
+ "learning_rate": 0.0019228976063615497,
+ "loss": 4.27017879486084,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.008840068243443966,
+ "learning_rate": 0.0019216157791038834,
+ "loss": 4.28773307800293,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.009512359276413918,
+ "learning_rate": 0.001920333624243228,
+ "loss": 4.266008377075195,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.009946667589247227,
+ "learning_rate": 0.001919051142812593,
+ "loss": 4.279062271118164,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.00899894442409277,
+ "learning_rate": 0.0019177683358452505,
+ "loss": 4.277679443359375,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.00909668579697609,
+ "learning_rate": 0.0019164852043747338,
+ "loss": 4.311397552490234,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.009787327609956264,
+ "learning_rate": 0.0019152017494348408,
+ "loss": 4.30875825881958,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.008983899839222431,
+ "learning_rate": 0.001913917972059626,
+ "loss": 4.259649276733398,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.008411582559347153,
+ "learning_rate": 0.0019126338732834071,
+ "loss": 4.259811878204346,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.009162918664515018,
+ "learning_rate": 0.0019113494541407579,
+ "loss": 4.271515846252441,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.011304561980068684,
+ "learning_rate": 0.0019100647156665125,
+ "loss": 4.269217491149902,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.012483394704759121,
+ "learning_rate": 0.0019087796588957617,
+ "loss": 4.2949299812316895,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.012091380544006824,
+ "learning_rate": 0.0019074942848638522,
+ "loss": 4.289752006530762,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.010011567734181881,
+ "learning_rate": 0.0019062085946063869,
+ "loss": 4.294095993041992,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.008766217157244682,
+ "learning_rate": 0.0019049225891592226,
+ "loss": 4.256425857543945,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.012862488627433777,
+ "learning_rate": 0.001903636269558471,
+ "loss": 4.287847995758057,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.013917063362896442,
+ "learning_rate": 0.0019023496368404972,
+ "loss": 4.282784938812256,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.012129023671150208,
+ "learning_rate": 0.0019010626920419173,
+ "loss": 4.2486419677734375,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.010348426178097725,
+ "learning_rate": 0.0018997754361995986,
+ "loss": 4.289836883544922,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.011821693740785122,
+ "learning_rate": 0.001898487870350662,
+ "loss": 4.214344024658203,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.013136724010109901,
+ "learning_rate": 0.0018971999955324742,
+ "loss": 4.24208402633667,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.013173168525099754,
+ "learning_rate": 0.0018959118127826538,
+ "loss": 4.270164489746094,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.011018159799277782,
+ "learning_rate": 0.0018946233231390656,
+ "loss": 4.2571916580200195,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.01018658559769392,
+ "learning_rate": 0.0018933345276398227,
+ "loss": 4.283267498016357,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.009412688203155994,
+ "learning_rate": 0.0018920454273232848,
+ "loss": 4.272198677062988,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.007799267303198576,
+ "learning_rate": 0.0018907560232280557,
+ "loss": 4.260125637054443,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.008747464045882225,
+ "learning_rate": 0.0018894663163929863,
+ "loss": 4.2539215087890625,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.008003434166312218,
+ "learning_rate": 0.0018881763078571689,
+ "loss": 4.245024681091309,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.007593846879899502,
+ "learning_rate": 0.0018868859986599415,
+ "loss": 4.237929344177246,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.008099420927464962,
+ "learning_rate": 0.0018855953898408815,
+ "loss": 4.2789106369018555,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.008270486257970333,
+ "learning_rate": 0.0018843044824398097,
+ "loss": 4.263522148132324,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.009260379709303379,
+ "learning_rate": 0.0018830132774967866,
+ "loss": 4.275835037231445,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.009289492852985859,
+ "learning_rate": 0.0018817217760521132,
+ "loss": 4.313778400421143,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.009493921883404255,
+ "learning_rate": 0.0018804299791463281,
+ "loss": 4.303672790527344,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.009020023047924042,
+ "learning_rate": 0.0018791378878202105,
+ "loss": 4.245473384857178,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.009446556679904461,
+ "learning_rate": 0.0018778455031147724,
+ "loss": 4.267327785491943,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.009499974548816681,
+ "learning_rate": 0.0018765528260712672,
+ "loss": 4.301205635070801,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.008399393409490585,
+ "learning_rate": 0.00187525985773118,
+ "loss": 4.307802677154541,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.009519701823592186,
+ "learning_rate": 0.0018739665991362327,
+ "loss": 4.284796714782715,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.01125232595950365,
+ "learning_rate": 0.0018726730513283804,
+ "loss": 4.285930633544922,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.011608537286520004,
+ "learning_rate": 0.00187137921534981,
+ "loss": 4.257689476013184,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.01064533181488514,
+ "learning_rate": 0.001870085092242944,
+ "loss": 4.249063968658447,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.009854599833488464,
+ "learning_rate": 0.0018687906830504321,
+ "loss": 4.255159378051758,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.01121739111840725,
+ "learning_rate": 0.0018674959888151576,
+ "loss": 4.287777900695801,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.5675368309021,
+ "eval_runtime": 51.8555,
+ "eval_samples_per_second": 47.092,
+ "eval_steps_per_second": 1.485,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.0104464590549469,
+ "learning_rate": 0.0018662010105802314,
+ "loss": 4.206080436706543,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.012800181284546852,
+ "learning_rate": 0.0018649057493889945,
+ "loss": 4.213550567626953,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.013382901437580585,
+ "learning_rate": 0.001863610206285015,
+ "loss": 4.190858840942383,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.014347050338983536,
+ "learning_rate": 0.0018623143823120894,
+ "loss": 4.22884464263916,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.01347152329981327,
+ "learning_rate": 0.001861018278514239,
+ "loss": 4.215579032897949,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.012786377221345901,
+ "learning_rate": 0.0018597218959357112,
+ "loss": 4.174999237060547,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.014137202873826027,
+ "learning_rate": 0.001858425235620979,
+ "loss": 4.21709680557251,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.013864399865269661,
+ "learning_rate": 0.001857128298614737,
+ "loss": 4.2375030517578125,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.010397658683359623,
+ "learning_rate": 0.0018558310859619048,
+ "loss": 4.231938362121582,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.00888819433748722,
+ "learning_rate": 0.0018545335987076223,
+ "loss": 4.195562839508057,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.010698067955672741,
+ "learning_rate": 0.001853235837897252,
+ "loss": 4.22847318649292,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.00987589918076992,
+ "learning_rate": 0.0018519378045763766,
+ "loss": 4.2147040367126465,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.011451782658696175,
+ "learning_rate": 0.001850639499790798,
+ "loss": 4.179263114929199,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.011224704794585705,
+ "learning_rate": 0.0018493409245865368,
+ "loss": 4.23030948638916,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.00996861606836319,
+ "learning_rate": 0.0018480420800098313,
+ "loss": 4.194283962249756,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.012143119238317013,
+ "learning_rate": 0.0018467429671071374,
+ "loss": 4.221704959869385,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.013126783072948456,
+ "learning_rate": 0.0018454435869251268,
+ "loss": 4.187748908996582,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.013031953014433384,
+ "learning_rate": 0.001844143940510687,
+ "loss": 4.248250961303711,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.012868993915617466,
+ "learning_rate": 0.001842844028910919,
+ "loss": 4.20161247253418,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.010102041065692902,
+ "learning_rate": 0.0018415438531731395,
+ "loss": 4.213654518127441,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.009599820710718632,
+ "learning_rate": 0.0018402434143448747,
+ "loss": 4.206313133239746,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.011536866426467896,
+ "learning_rate": 0.0018389427134738658,
+ "loss": 4.237072944641113,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.009558089077472687,
+ "learning_rate": 0.0018376417516080636,
+ "loss": 4.182195663452148,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01056219544261694,
+ "learning_rate": 0.0018363405297956302,
+ "loss": 4.232585906982422,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.012673584744334221,
+ "learning_rate": 0.0018350390490849354,
+ "loss": 4.228830337524414,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.011215884238481522,
+ "learning_rate": 0.0018337373105245598,
+ "loss": 4.216056823730469,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.011281915940344334,
+ "learning_rate": 0.00183243531516329,
+ "loss": 4.233404159545898,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.010381227359175682,
+ "learning_rate": 0.0018311330640501204,
+ "loss": 4.182013034820557,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.00900198519229889,
+ "learning_rate": 0.0018298305582342513,
+ "loss": 4.215488433837891,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.010198601521551609,
+ "learning_rate": 0.0018285277987650876,
+ "loss": 4.19389533996582,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.009965402074158192,
+ "learning_rate": 0.0018272247866922401,
+ "loss": 4.218005657196045,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.009299935773015022,
+ "learning_rate": 0.0018259215230655202,
+ "loss": 4.223684310913086,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.008977951481938362,
+ "learning_rate": 0.0018246180089349462,
+ "loss": 4.215778350830078,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.009261307306587696,
+ "learning_rate": 0.0018233142453507344,
+ "loss": 4.215066909790039,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.008322088047862053,
+ "learning_rate": 0.0018220102333633044,
+ "loss": 4.2188005447387695,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.008716144599020481,
+ "learning_rate": 0.0018207059740232739,
+ "loss": 4.187867641448975,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.008897020481526852,
+ "learning_rate": 0.0018194014683814625,
+ "loss": 4.218034267425537,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.007298174314200878,
+ "learning_rate": 0.0018180967174888857,
+ "loss": 4.2017011642456055,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.008491789922118187,
+ "learning_rate": 0.0018167917223967582,
+ "loss": 4.239873886108398,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.008413378149271011,
+ "learning_rate": 0.0018154864841564908,
+ "loss": 4.234771251678467,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.008141471073031425,
+ "learning_rate": 0.0018141810038196909,
+ "loss": 4.185944080352783,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.009480944834649563,
+ "learning_rate": 0.0018128752824381603,
+ "loss": 4.206586837768555,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.0106047997251153,
+ "learning_rate": 0.001811569321063894,
+ "loss": 4.235287666320801,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.010663514956831932,
+ "learning_rate": 0.0018102631207490838,
+ "loss": 4.204899311065674,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.011473523452877998,
+ "learning_rate": 0.001808956682546109,
+ "loss": 4.253668308258057,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01165992021560669,
+ "learning_rate": 0.0018076500075075456,
+ "loss": 4.209845542907715,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.011692410334944725,
+ "learning_rate": 0.0018063430966861564,
+ "loss": 4.250368595123291,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01061506848782301,
+ "learning_rate": 0.0018050359511348965,
+ "loss": 4.193821907043457,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.010510941036045551,
+ "learning_rate": 0.0018037285719069097,
+ "loss": 4.254327774047852,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.011587041430175304,
+ "learning_rate": 0.001802420960055528,
+ "loss": 4.2838945388793945,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.01257420051842928,
+ "learning_rate": 0.0018011131166342693,
+ "loss": 4.258241653442383,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.00936233438551426,
+ "learning_rate": 0.0017998050426968407,
+ "loss": 4.201854705810547,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.010093127377331257,
+ "learning_rate": 0.0017984967392971338,
+ "loss": 4.193051338195801,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.010032009333372116,
+ "learning_rate": 0.0017971882074892236,
+ "loss": 4.230212211608887,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.009455587714910507,
+ "learning_rate": 0.0017958794483273726,
+ "loss": 4.2157673835754395,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.009549238719046116,
+ "learning_rate": 0.001794570462866023,
+ "loss": 4.260272026062012,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.011059699580073357,
+ "learning_rate": 0.0017932612521598013,
+ "loss": 4.214818000793457,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.011653627268970013,
+ "learning_rate": 0.0017919518172635146,
+ "loss": 4.205112457275391,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.012438314035534859,
+ "learning_rate": 0.001790642159232151,
+ "loss": 4.179045677185059,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.011234673671424389,
+ "learning_rate": 0.0017893322791208784,
+ "loss": 4.239317417144775,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.010923845693469048,
+ "learning_rate": 0.0017880221779850434,
+ "loss": 4.230571269989014,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.009219309315085411,
+ "learning_rate": 0.001786711856880171,
+ "loss": 4.229068756103516,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.009849398396909237,
+ "learning_rate": 0.0017854013168619638,
+ "loss": 4.254772186279297,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.01117763202637434,
+ "learning_rate": 0.001784090558986299,
+ "loss": 4.252647399902344,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.010664612054824829,
+ "learning_rate": 0.0017827795843092316,
+ "loss": 4.204596519470215,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.00847950391471386,
+ "learning_rate": 0.0017814683938869895,
+ "loss": 4.211061954498291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.007305726408958435,
+ "learning_rate": 0.0017801569887759753,
+ "loss": 4.181221008300781,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.007970666512846947,
+ "learning_rate": 0.0017788453700327653,
+ "loss": 4.237855434417725,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.008873614482581615,
+ "learning_rate": 0.0017775335387141054,
+ "loss": 4.229188919067383,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.009691286832094193,
+ "learning_rate": 0.0017762214958769156,
+ "loss": 4.205491542816162,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.009360672906041145,
+ "learning_rate": 0.0017749092425782852,
+ "loss": 4.233466625213623,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.009895884431898594,
+ "learning_rate": 0.0017735967798754723,
+ "loss": 4.187763214111328,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.535672187805176,
+ "eval_runtime": 53.1736,
+ "eval_samples_per_second": 45.925,
+ "eval_steps_per_second": 1.448,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.00966129545122385,
+ "learning_rate": 0.0017722841088259045,
+ "loss": 4.1451735496521,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.010733380913734436,
+ "learning_rate": 0.0017709712304871774,
+ "loss": 4.1616363525390625,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.0117328567430377,
+ "learning_rate": 0.0017696581459170535,
+ "loss": 4.126331806182861,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.012640298344194889,
+ "learning_rate": 0.0017683448561734607,
+ "loss": 4.163730621337891,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.013074076734483242,
+ "learning_rate": 0.0017670313623144934,
+ "loss": 4.199502944946289,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.01132719125598669,
+ "learning_rate": 0.0017657176653984097,
+ "loss": 4.143380165100098,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.010008333250880241,
+ "learning_rate": 0.0017644037664836319,
+ "loss": 4.170320510864258,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.011090578511357307,
+ "learning_rate": 0.0017630896666287438,
+ "loss": 4.1571478843688965,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.011047456413507462,
+ "learning_rate": 0.0017617753668924924,
+ "loss": 4.101892948150635,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.010098653845489025,
+ "learning_rate": 0.0017604608683337857,
+ "loss": 4.154077529907227,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.010499860160052776,
+ "learning_rate": 0.00175914617201169,
+ "loss": 4.116868019104004,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.010939318686723709,
+ "learning_rate": 0.0017578312789854338,
+ "loss": 4.159144878387451,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.010622999630868435,
+ "learning_rate": 0.0017565161903144023,
+ "loss": 4.171411991119385,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.0108700692653656,
+ "learning_rate": 0.0017552009070581386,
+ "loss": 4.100748062133789,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.009794988669455051,
+ "learning_rate": 0.0017538854302763423,
+ "loss": 4.185161590576172,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.007750354241579771,
+ "learning_rate": 0.0017525697610288702,
+ "loss": 4.149083137512207,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.009923134930431843,
+ "learning_rate": 0.001751253900375732,
+ "loss": 4.205166816711426,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.010811141692101955,
+ "learning_rate": 0.0017499378493770936,
+ "loss": 4.131865978240967,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.010868688113987446,
+ "learning_rate": 0.0017486216090932733,
+ "loss": 4.148791790008545,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.011224126443266869,
+ "learning_rate": 0.001747305180584743,
+ "loss": 4.152503490447998,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.013262554071843624,
+ "learning_rate": 0.001745988564912124,
+ "loss": 4.164926528930664,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.013089599087834358,
+ "learning_rate": 0.001744671763136191,
+ "loss": 4.1626458168029785,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.01327646616846323,
+ "learning_rate": 0.0017433547763178665,
+ "loss": 4.189356803894043,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.014539042487740517,
+ "learning_rate": 0.0017420376055182233,
+ "loss": 4.176387786865234,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.013269098475575447,
+ "learning_rate": 0.001740720251798482,
+ "loss": 4.180716514587402,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.010922391898930073,
+ "learning_rate": 0.0017394027162200105,
+ "loss": 4.163928031921387,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.013355271890759468,
+ "learning_rate": 0.0017380849998443243,
+ "loss": 4.150012016296387,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.014149201102554798,
+ "learning_rate": 0.0017367671037330828,
+ "loss": 4.173713207244873,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01354879979044199,
+ "learning_rate": 0.0017354490289480918,
+ "loss": 4.150563716888428,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.013788704760372639,
+ "learning_rate": 0.0017341307765512998,
+ "loss": 4.213858604431152,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.011597222648561,
+ "learning_rate": 0.0017328123476047994,
+ "loss": 4.166989326477051,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.01055675558745861,
+ "learning_rate": 0.001731493743170824,
+ "loss": 4.172226428985596,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.011600397527217865,
+ "learning_rate": 0.0017301749643117506,
+ "loss": 4.1656951904296875,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.01114651933312416,
+ "learning_rate": 0.0017288560120900943,
+ "loss": 4.189825057983398,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.009811583906412125,
+ "learning_rate": 0.0017275368875685126,
+ "loss": 4.181164741516113,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.009758968837559223,
+ "learning_rate": 0.0017262175918097992,
+ "loss": 4.163188934326172,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.00999707542359829,
+ "learning_rate": 0.001724898125876887,
+ "loss": 4.184117317199707,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.009528180584311485,
+ "learning_rate": 0.001723578490832846,
+ "loss": 4.16407585144043,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.009717358276247978,
+ "learning_rate": 0.0017222586877408826,
+ "loss": 4.209892272949219,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.010909050703048706,
+ "learning_rate": 0.0017209387176643377,
+ "loss": 4.19977331161499,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.011234750039875507,
+ "learning_rate": 0.0017196185816666875,
+ "loss": 4.19363260269165,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.010300721041858196,
+ "learning_rate": 0.0017182982808115425,
+ "loss": 4.142427921295166,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.00941646471619606,
+ "learning_rate": 0.001716977816162644,
+ "loss": 4.159372329711914,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.009304054081439972,
+ "learning_rate": 0.0017156571887838679,
+ "loss": 4.207922458648682,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.009496180340647697,
+ "learning_rate": 0.0017143363997392184,
+ "loss": 4.137894630432129,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.009257436729967594,
+ "learning_rate": 0.0017130154500928326,
+ "loss": 4.198753833770752,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.00900586973875761,
+ "learning_rate": 0.0017116943409089746,
+ "loss": 4.160684585571289,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.009398494847118855,
+ "learning_rate": 0.00171037307325204,
+ "loss": 4.1819353103637695,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.008633013814687729,
+ "learning_rate": 0.0017090516481865483,
+ "loss": 4.176833152770996,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.00898425467312336,
+ "learning_rate": 0.0017077300667771484,
+ "loss": 4.190657138824463,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.008182398043572903,
+ "learning_rate": 0.0017064083300886155,
+ "loss": 4.173429489135742,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.009589143097400665,
+ "learning_rate": 0.0017050864391858478,
+ "loss": 4.1559014320373535,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.010304383002221584,
+ "learning_rate": 0.0017037643951338698,
+ "loss": 4.155453681945801,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.009943748824298382,
+ "learning_rate": 0.0017024421989978275,
+ "loss": 4.119734287261963,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.009238033555448055,
+ "learning_rate": 0.001701119851842992,
+ "loss": 4.188974857330322,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.008058690465986729,
+ "learning_rate": 0.0016997973547347531,
+ "loss": 4.19934606552124,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.008485953323543072,
+ "learning_rate": 0.0016984747087386233,
+ "loss": 4.165615558624268,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.008476605638861656,
+ "learning_rate": 0.0016971519149202347,
+ "loss": 4.21021842956543,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.008588708005845547,
+ "learning_rate": 0.0016958289743453385,
+ "loss": 4.208064556121826,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.009223178029060364,
+ "learning_rate": 0.0016945058880798037,
+ "loss": 4.158413887023926,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.009196167811751366,
+ "learning_rate": 0.0016931826571896166,
+ "loss": 4.203187465667725,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.009488921612501144,
+ "learning_rate": 0.001691859282740882,
+ "loss": 4.178227424621582,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.009253161028027534,
+ "learning_rate": 0.0016905357657998167,
+ "loss": 4.131230354309082,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.010069048963487148,
+ "learning_rate": 0.0016892121074327559,
+ "loss": 4.210734844207764,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.009945095516741276,
+ "learning_rate": 0.001687888308706147,
+ "loss": 4.184042930603027,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.010483640246093273,
+ "learning_rate": 0.0016865643706865505,
+ "loss": 4.156919479370117,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.011560715734958649,
+ "learning_rate": 0.0016852402944406391,
+ "loss": 4.185520172119141,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.010001277551054955,
+ "learning_rate": 0.0016839160810351978,
+ "loss": 4.226828575134277,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.010335934348404408,
+ "learning_rate": 0.0016825917315371202,
+ "loss": 4.183531284332275,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.011322776786983013,
+ "learning_rate": 0.001681267247013412,
+ "loss": 4.179635047912598,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.009663194417953491,
+ "learning_rate": 0.001679942628531186,
+ "loss": 4.165654182434082,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.010590246878564358,
+ "learning_rate": 0.0016786178771576643,
+ "loss": 4.1539459228515625,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.511386394500732,
+ "eval_runtime": 51.7751,
+ "eval_samples_per_second": 47.166,
+ "eval_steps_per_second": 1.487,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.01254354789853096,
+ "learning_rate": 0.0016772929939601728,
+ "loss": 4.104002475738525,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.01300987508147955,
+ "learning_rate": 0.0016759679800061483,
+ "loss": 4.127395153045654,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.013325816951692104,
+ "learning_rate": 0.0016746428363631299,
+ "loss": 4.131229400634766,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.014302870258688927,
+ "learning_rate": 0.001673317564098761,
+ "loss": 4.079358100891113,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.013776068575680256,
+ "learning_rate": 0.0016719921642807905,
+ "loss": 4.132065773010254,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.010788560844957829,
+ "learning_rate": 0.0016706666379770689,
+ "loss": 4.134626388549805,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.009766630828380585,
+ "learning_rate": 0.0016693409862555493,
+ "loss": 4.1645426750183105,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.011760029010474682,
+ "learning_rate": 0.0016680152101842842,
+ "loss": 4.110664367675781,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.011047836393117905,
+ "learning_rate": 0.0016666893108314288,
+ "loss": 4.129222869873047,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.009989139623939991,
+ "learning_rate": 0.0016653632892652344,
+ "loss": 4.076835632324219,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.010033783502876759,
+ "learning_rate": 0.001664037146554054,
+ "loss": 4.11555290222168,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.00996427796781063,
+ "learning_rate": 0.001662710883766337,
+ "loss": 4.131463527679443,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.01044023409485817,
+ "learning_rate": 0.0016613845019706288,
+ "loss": 4.098750114440918,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.010033872909843922,
+ "learning_rate": 0.0016600580022355712,
+ "loss": 4.161447048187256,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.009329567663371563,
+ "learning_rate": 0.0016587313856299012,
+ "loss": 4.146491527557373,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.010061589069664478,
+ "learning_rate": 0.0016574046532224504,
+ "loss": 4.107517719268799,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.009491910226643085,
+ "learning_rate": 0.0016560778060821424,
+ "loss": 4.098226547241211,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.008529345504939556,
+ "learning_rate": 0.0016547508452779943,
+ "loss": 4.1247711181640625,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.00936342403292656,
+ "learning_rate": 0.0016534237718791136,
+ "loss": 4.120573043823242,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.010255641303956509,
+ "learning_rate": 0.0016520965869547008,
+ "loss": 4.089067459106445,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.011638758704066277,
+ "learning_rate": 0.001650769291574044,
+ "loss": 4.074847221374512,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.014707857742905617,
+ "learning_rate": 0.0016494418868065214,
+ "loss": 4.153377532958984,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.01650111749768257,
+ "learning_rate": 0.0016481143737215986,
+ "loss": 4.134062767028809,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.014833372086286545,
+ "learning_rate": 0.00164678675338883,
+ "loss": 4.1344828605651855,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.012257886119186878,
+ "learning_rate": 0.0016454590268778538,
+ "loss": 4.098813533782959,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.013508322648704052,
+ "learning_rate": 0.0016441311952583959,
+ "loss": 4.106297492980957,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.01256900280714035,
+ "learning_rate": 0.001642803259600267,
+ "loss": 4.143631935119629,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.009885406121611595,
+ "learning_rate": 0.00164147522097336,
+ "loss": 4.120885848999023,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.012004674412310123,
+ "learning_rate": 0.0016401470804476518,
+ "loss": 4.089243412017822,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01224028691649437,
+ "learning_rate": 0.0016388188390932012,
+ "loss": 4.077765464782715,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.012511290609836578,
+ "learning_rate": 0.001637490497980149,
+ "loss": 4.161346435546875,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.010701716877520084,
+ "learning_rate": 0.0016361620581787143,
+ "loss": 4.10698938369751,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.010713113471865654,
+ "learning_rate": 0.001634833520759198,
+ "loss": 4.114283561706543,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.011018907651305199,
+ "learning_rate": 0.001633504886791978,
+ "loss": 4.134284019470215,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.011279584839940071,
+ "learning_rate": 0.0016321761573475117,
+ "loss": 4.1456804275512695,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.010642931796610355,
+ "learning_rate": 0.001630847333496332,
+ "loss": 4.160918235778809,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.011069678701460361,
+ "learning_rate": 0.0016295184163090482,
+ "loss": 4.136838912963867,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.011007166467607021,
+ "learning_rate": 0.0016281894068563443,
+ "loss": 4.129153728485107,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.010055703110992908,
+ "learning_rate": 0.0016268603062089802,
+ "loss": 4.138665676116943,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.009839230217039585,
+ "learning_rate": 0.0016255311154377875,
+ "loss": 4.1366496086120605,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.011191432364284992,
+ "learning_rate": 0.0016242018356136714,
+ "loss": 4.129880905151367,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.012130341492593288,
+ "learning_rate": 0.0016228724678076095,
+ "loss": 4.1425981521606445,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.012707795947790146,
+ "learning_rate": 0.0016215430130906482,
+ "loss": 4.115452766418457,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.012707809917628765,
+ "learning_rate": 0.001620213472533906,
+ "loss": 4.122758865356445,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.01182397548109293,
+ "learning_rate": 0.001618883847208569,
+ "loss": 4.140275955200195,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.010865340940654278,
+ "learning_rate": 0.001617554138185893,
+ "loss": 4.137759208679199,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.010973592288792133,
+ "learning_rate": 0.0016162243465371998,
+ "loss": 4.138947010040283,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.011247226968407631,
+ "learning_rate": 0.001614894473333879,
+ "loss": 4.115516662597656,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.010837487876415253,
+ "learning_rate": 0.0016135645196473853,
+ "loss": 4.160427093505859,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.009096396155655384,
+ "learning_rate": 0.0016122344865492385,
+ "loss": 4.1200032234191895,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.010671247728168964,
+ "learning_rate": 0.0016109043751110223,
+ "loss": 4.097919464111328,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.0107562355697155,
+ "learning_rate": 0.0016095741864043831,
+ "loss": 4.096290588378906,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.008909327909350395,
+ "learning_rate": 0.0016082439215010305,
+ "loss": 4.171017646789551,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.010305124334990978,
+ "learning_rate": 0.0016069135814727343,
+ "loss": 4.1450276374816895,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.011235613375902176,
+ "learning_rate": 0.0016055831673913255,
+ "loss": 4.134704113006592,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.00825632456690073,
+ "learning_rate": 0.0016042526803286949,
+ "loss": 4.107443332672119,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.009056378155946732,
+ "learning_rate": 0.0016029221213567921,
+ "loss": 4.132656097412109,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.008835164830088615,
+ "learning_rate": 0.001601591491547624,
+ "loss": 4.124729633331299,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.00915333442389965,
+ "learning_rate": 0.0016002607919732557,
+ "loss": 4.119500160217285,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.008532539941370487,
+ "learning_rate": 0.0015989300237058066,
+ "loss": 4.141661643981934,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.009952311404049397,
+ "learning_rate": 0.0015975991878174533,
+ "loss": 4.140929222106934,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.008920297026634216,
+ "learning_rate": 0.0015962682853804268,
+ "loss": 4.155877113342285,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.008770049549639225,
+ "learning_rate": 0.0015949373174670104,
+ "loss": 4.126394271850586,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.0078084105625748634,
+ "learning_rate": 0.0015936062851495411,
+ "loss": 4.128876209259033,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.00816641841083765,
+ "learning_rate": 0.0015922751895004078,
+ "loss": 4.137410640716553,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.008895459584891796,
+ "learning_rate": 0.0015909440315920505,
+ "loss": 4.07957649230957,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.008551528677344322,
+ "learning_rate": 0.001589612812496958,
+ "loss": 4.104213714599609,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.008325488306581974,
+ "learning_rate": 0.0015882815332876707,
+ "loss": 4.1584792137146,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.009341313503682613,
+ "learning_rate": 0.0015869501950367744,
+ "loss": 4.1417436599731445,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.00999050959944725,
+ "learning_rate": 0.0015856187988169067,
+ "loss": 4.136049270629883,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.008889932185411453,
+ "learning_rate": 0.0015842873457007484,
+ "loss": 4.16257381439209,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.009109573438763618,
+ "learning_rate": 0.0015829558367610268,
+ "loss": 4.171441555023193,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.485942840576172,
+ "eval_runtime": 52.1051,
+ "eval_samples_per_second": 46.867,
+ "eval_steps_per_second": 1.478,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.009881424717605114,
+ "learning_rate": 0.001581624273070515,
+ "loss": 4.059608459472656,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.011156721040606499,
+ "learning_rate": 0.0015802926557020294,
+ "loss": 4.047014236450195,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.012326992116868496,
+ "learning_rate": 0.0015789609857284309,
+ "loss": 4.06541109085083,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.01332208700478077,
+ "learning_rate": 0.0015776292642226213,
+ "loss": 4.042714595794678,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.012720128521323204,
+ "learning_rate": 0.0015762974922575446,
+ "loss": 4.098389625549316,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.010723215527832508,
+ "learning_rate": 0.0015749656709061854,
+ "loss": 4.059185981750488,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.010605281218886375,
+ "learning_rate": 0.001573633801241569,
+ "loss": 4.099538326263428,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.01101530622690916,
+ "learning_rate": 0.0015723018843367576,
+ "loss": 4.095674514770508,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.01135838869959116,
+ "learning_rate": 0.0015709699212648532,
+ "loss": 4.070535659790039,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.012119271792471409,
+ "learning_rate": 0.0015696379130989936,
+ "loss": 4.060605049133301,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.012783961370587349,
+ "learning_rate": 0.001568305860912355,
+ "loss": 4.075931072235107,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.011486220173537731,
+ "learning_rate": 0.0015669737657781461,
+ "loss": 4.075855255126953,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.01037552859634161,
+ "learning_rate": 0.0015656416287696136,
+ "loss": 4.087729454040527,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.009802743792533875,
+ "learning_rate": 0.0015643094509600354,
+ "loss": 4.0855607986450195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.010804438032209873,
+ "learning_rate": 0.001562977233422723,
+ "loss": 4.064528465270996,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.009505728259682655,
+ "learning_rate": 0.0015616449772310206,
+ "loss": 4.036775588989258,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.010634485632181168,
+ "learning_rate": 0.0015603126834583024,
+ "loss": 4.069923400878906,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.011392096057534218,
+ "learning_rate": 0.0015589803531779736,
+ "loss": 4.0812177658081055,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.011851722374558449,
+ "learning_rate": 0.0015576479874634683,
+ "loss": 4.09183406829834,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.011205444112420082,
+ "learning_rate": 0.0015563155873882502,
+ "loss": 4.07825231552124,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.009676349349319935,
+ "learning_rate": 0.0015549831540258093,
+ "loss": 4.052826404571533,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.009376165457069874,
+ "learning_rate": 0.0015536506884496638,
+ "loss": 4.02100133895874,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.00936955213546753,
+ "learning_rate": 0.0015523181917333564,
+ "loss": 4.049526214599609,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.00937784556299448,
+ "learning_rate": 0.0015509856649504564,
+ "loss": 4.096356391906738,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.010014008730649948,
+ "learning_rate": 0.0015496531091745555,
+ "loss": 4.03254508972168,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.009909634478390217,
+ "learning_rate": 0.0015483205254792705,
+ "loss": 4.067630767822266,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.009319501928985119,
+ "learning_rate": 0.0015469879149382399,
+ "loss": 4.094167709350586,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.008634721860289574,
+ "learning_rate": 0.0015456552786251244,
+ "loss": 4.122404098510742,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.009412454441189766,
+ "learning_rate": 0.0015443226176136049,
+ "loss": 4.081546783447266,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.010102108120918274,
+ "learning_rate": 0.0015429899329773814,
+ "loss": 4.049839019775391,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.010921654291450977,
+ "learning_rate": 0.0015416572257901746,
+ "loss": 4.097097396850586,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.010607713833451271,
+ "learning_rate": 0.0015403244971257226,
+ "loss": 4.084481239318848,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.011887960135936737,
+ "learning_rate": 0.0015389917480577806,
+ "loss": 4.074968338012695,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01089888159185648,
+ "learning_rate": 0.0015376589796601203,
+ "loss": 4.062132835388184,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.009934183210134506,
+ "learning_rate": 0.0015363261930065298,
+ "loss": 4.128092288970947,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.00878534372895956,
+ "learning_rate": 0.0015349933891708105,
+ "loss": 4.069051742553711,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.009928155690431595,
+ "learning_rate": 0.0015336605692267788,
+ "loss": 4.073912620544434,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.011091411113739014,
+ "learning_rate": 0.0015323277342482637,
+ "loss": 4.125944137573242,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.010314172133803368,
+ "learning_rate": 0.0015309948853091055,
+ "loss": 4.062620162963867,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.010423967614769936,
+ "learning_rate": 0.001529662023483158,
+ "loss": 4.101737976074219,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.009822997264564037,
+ "learning_rate": 0.0015283291498442817,
+ "loss": 4.064520359039307,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.009757095016539097,
+ "learning_rate": 0.001526996265466351,
+ "loss": 4.100624084472656,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.008650688454508781,
+ "learning_rate": 0.001525663371423246,
+ "loss": 4.109685897827148,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.007378234062343836,
+ "learning_rate": 0.0015243304687888553,
+ "loss": 4.044726848602295,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.008194336667656898,
+ "learning_rate": 0.001522997558637074,
+ "loss": 4.097923278808594,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.009764892049133778,
+ "learning_rate": 0.0015216646420418048,
+ "loss": 4.102148056030273,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.009242868050932884,
+ "learning_rate": 0.001520331720076953,
+ "loss": 4.068563461303711,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.010409324429929256,
+ "learning_rate": 0.001518998793816431,
+ "loss": 4.065944194793701,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.009215258061885834,
+ "learning_rate": 0.0015176658643341527,
+ "loss": 4.10223388671875,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.008065393194556236,
+ "learning_rate": 0.0015163329327040366,
+ "loss": 4.099432945251465,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.008498288691043854,
+ "learning_rate": 0.001515,
+ "loss": 4.109854698181152,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.008919150568544865,
+ "learning_rate": 0.0015136670672959633,
+ "loss": 4.128247261047363,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.009104429744184017,
+ "learning_rate": 0.0015123341356658474,
+ "loss": 4.068622589111328,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.008558365516364574,
+ "learning_rate": 0.0015110012061835692,
+ "loss": 4.114214897155762,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.007865076884627342,
+ "learning_rate": 0.0015096682799230477,
+ "loss": 4.11155366897583,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.007848534733057022,
+ "learning_rate": 0.0015083353579581958,
+ "loss": 4.045694351196289,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.008238030597567558,
+ "learning_rate": 0.0015070024413629264,
+ "loss": 4.093930721282959,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.007536398712545633,
+ "learning_rate": 0.0015056695312111448,
+ "loss": 4.106235504150391,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.008156892843544483,
+ "learning_rate": 0.001504336628576754,
+ "loss": 4.13060998916626,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.008144439198076725,
+ "learning_rate": 0.0015030037345336487,
+ "loss": 4.091304779052734,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.007757358253002167,
+ "learning_rate": 0.001501670850155718,
+ "loss": 4.069276809692383,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.00860623549669981,
+ "learning_rate": 0.0015003379765168425,
+ "loss": 4.107658386230469,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.008614457212388515,
+ "learning_rate": 0.0014990051146908949,
+ "loss": 4.106705665588379,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.009120592847466469,
+ "learning_rate": 0.001497672265751737,
+ "loss": 4.0891923904418945,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.009446980431675911,
+ "learning_rate": 0.0014963394307732213,
+ "loss": 4.098847389221191,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.010494766756892204,
+ "learning_rate": 0.0014950066108291894,
+ "loss": 4.115325450897217,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.010057658888399601,
+ "learning_rate": 0.0014936738069934706,
+ "loss": 4.099984169006348,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.009071771055459976,
+ "learning_rate": 0.0014923410203398796,
+ "loss": 4.091543674468994,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.009937075898051262,
+ "learning_rate": 0.00149100825194222,
+ "loss": 4.086492538452148,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.010222986340522766,
+ "learning_rate": 0.0014896755028742778,
+ "loss": 4.063841819763184,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.010513762012124062,
+ "learning_rate": 0.001488342774209826,
+ "loss": 4.11991548538208,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.009530222974717617,
+ "learning_rate": 0.0014870100670226192,
+ "loss": 4.043334007263184,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.467459678649902,
+ "eval_runtime": 54.1628,
+ "eval_samples_per_second": 45.086,
+ "eval_steps_per_second": 1.422,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.010274187661707401,
+ "learning_rate": 0.0014856773823863955,
+ "loss": 4.049561500549316,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.011448442004621029,
+ "learning_rate": 0.001484344721374876,
+ "loss": 4.019484996795654,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.011361192911863327,
+ "learning_rate": 0.0014830120850617599,
+ "loss": 4.0429887771606445,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.010250156745314598,
+ "learning_rate": 0.0014816794745207299,
+ "loss": 4.012682914733887,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.010024400427937508,
+ "learning_rate": 0.0014803468908254449,
+ "loss": 3.979409694671631,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.01153535582125187,
+ "learning_rate": 0.0014790143350495446,
+ "loss": 4.045780181884766,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.012632817961275578,
+ "learning_rate": 0.001477681808266644,
+ "loss": 4.057641983032227,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.013758293353021145,
+ "learning_rate": 0.0014763493115503366,
+ "loss": 4.061607360839844,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.014041583053767681,
+ "learning_rate": 0.001475016845974191,
+ "loss": 4.035029411315918,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.013629456982016563,
+ "learning_rate": 0.0014736844126117497,
+ "loss": 4.062169075012207,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.011023213155567646,
+ "learning_rate": 0.0014723520125365323,
+ "loss": 4.02731990814209,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.01001750584691763,
+ "learning_rate": 0.0014710196468220266,
+ "loss": 4.019882678985596,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.008321646600961685,
+ "learning_rate": 0.001469687316541698,
+ "loss": 4.005446434020996,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.008935946971178055,
+ "learning_rate": 0.0014683550227689796,
+ "loss": 4.017102241516113,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.009137130342423916,
+ "learning_rate": 0.001467022766577277,
+ "loss": 4.024102210998535,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.00805022194981575,
+ "learning_rate": 0.001465690549039965,
+ "loss": 3.9935250282287598,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.009204058907926083,
+ "learning_rate": 0.0014643583712303865,
+ "loss": 4.042818546295166,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.009812815114855766,
+ "learning_rate": 0.001463026234221854,
+ "loss": 4.0252485275268555,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.009955782443284988,
+ "learning_rate": 0.0014616941390876454,
+ "loss": 3.9839768409729004,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.009374645538628101,
+ "learning_rate": 0.001460362086901007,
+ "loss": 4.040245056152344,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.010008739307522774,
+ "learning_rate": 0.0014590300787351471,
+ "loss": 4.0306596755981445,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.009014200419187546,
+ "learning_rate": 0.001457698115663243,
+ "loss": 4.028713226318359,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.008381523191928864,
+ "learning_rate": 0.0014563661987584313,
+ "loss": 4.024718284606934,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.008627948351204395,
+ "learning_rate": 0.0014550343290938143,
+ "loss": 4.02303409576416,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.010257705114781857,
+ "learning_rate": 0.0014537025077424556,
+ "loss": 4.061058044433594,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.010759209282696247,
+ "learning_rate": 0.0014523707357773788,
+ "loss": 4.064718246459961,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.009966139681637287,
+ "learning_rate": 0.0014510390142715697,
+ "loss": 3.9919447898864746,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.00997584406286478,
+ "learning_rate": 0.001449707344297971,
+ "loss": 4.0431599617004395,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.010846944525837898,
+ "learning_rate": 0.0014483757269294858,
+ "loss": 4.0222697257995605,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.00991352554410696,
+ "learning_rate": 0.0014470441632389735,
+ "loss": 4.059438705444336,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.008994397707283497,
+ "learning_rate": 0.0014457126542992518,
+ "loss": 4.021494388580322,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.009567267261445522,
+ "learning_rate": 0.0014443812011830934,
+ "loss": 4.070810317993164,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.009199011139571667,
+ "learning_rate": 0.0014430498049632255,
+ "loss": 4.020938873291016,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.010440350510179996,
+ "learning_rate": 0.0014417184667123303,
+ "loss": 4.062529563903809,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.009962482377886772,
+ "learning_rate": 0.0014403871875030424,
+ "loss": 4.0687150955200195,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.009389379993081093,
+ "learning_rate": 0.0014390559684079502,
+ "loss": 4.055185794830322,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.011469941586256027,
+ "learning_rate": 0.0014377248104995924,
+ "loss": 4.041023254394531,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.011034650728106499,
+ "learning_rate": 0.0014363937148504588,
+ "loss": 4.054254531860352,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.010651151649653912,
+ "learning_rate": 0.0014350626825329897,
+ "loss": 4.037753582000732,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.010747930034995079,
+ "learning_rate": 0.0014337317146195733,
+ "loss": 4.059938907623291,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.00987576600164175,
+ "learning_rate": 0.0014324008121825468,
+ "loss": 4.075366973876953,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.008504371158778667,
+ "learning_rate": 0.0014310699762941936,
+ "loss": 4.026755332946777,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.009445052593946457,
+ "learning_rate": 0.001429739208026745,
+ "loss": 4.034431457519531,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.010560023598372936,
+ "learning_rate": 0.0014284085084523763,
+ "loss": 4.101158618927002,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.009920580312609673,
+ "learning_rate": 0.0014270778786432078,
+ "loss": 4.041321754455566,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.009854268282651901,
+ "learning_rate": 0.0014257473196713053,
+ "loss": 4.049882411956787,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.009825135581195354,
+ "learning_rate": 0.0014244168326086745,
+ "loss": 4.067440986633301,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.010506784543395042,
+ "learning_rate": 0.0014230864185272663,
+ "loss": 4.042476654052734,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.009527767077088356,
+ "learning_rate": 0.0014217560784989696,
+ "loss": 4.055248260498047,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.01013826485723257,
+ "learning_rate": 0.001420425813595617,
+ "loss": 4.05246639251709,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.011020111851394176,
+ "learning_rate": 0.0014190956248889778,
+ "loss": 4.032987117767334,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.010647117160260677,
+ "learning_rate": 0.0014177655134507614,
+ "loss": 4.093589782714844,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.009484780952334404,
+ "learning_rate": 0.001416435480352615,
+ "loss": 4.017437934875488,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.0114199910312891,
+ "learning_rate": 0.0014151055266661211,
+ "loss": 4.080643653869629,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.011137664318084717,
+ "learning_rate": 0.0014137756534628007,
+ "loss": 4.069380760192871,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.008972682990133762,
+ "learning_rate": 0.0014124458618141073,
+ "loss": 4.051383018493652,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.010311643593013287,
+ "learning_rate": 0.0014111161527914317,
+ "loss": 4.048953533172607,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.01113453321158886,
+ "learning_rate": 0.0014097865274660944,
+ "loss": 4.063544273376465,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.01091703586280346,
+ "learning_rate": 0.0014084569869093518,
+ "loss": 4.006721496582031,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.010672449134290218,
+ "learning_rate": 0.0014071275321923906,
+ "loss": 4.020220756530762,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.009682527743279934,
+ "learning_rate": 0.001405798164386328,
+ "loss": 4.057307243347168,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.009808676317334175,
+ "learning_rate": 0.0014044688845622126,
+ "loss": 4.105412483215332,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.010351189412176609,
+ "learning_rate": 0.0014031396937910202,
+ "loss": 4.066454887390137,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.010038275271654129,
+ "learning_rate": 0.001401810593143656,
+ "loss": 4.056752681732178,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01016094908118248,
+ "learning_rate": 0.001400481583690952,
+ "loss": 4.097933292388916,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.008683547377586365,
+ "learning_rate": 0.0013991526665036679,
+ "loss": 4.0386152267456055,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.009556332603096962,
+ "learning_rate": 0.0013978238426524882,
+ "loss": 4.05835485458374,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.009597044438123703,
+ "learning_rate": 0.001396495113208022,
+ "loss": 4.052363872528076,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.010021139867603779,
+ "learning_rate": 0.0013951664792408026,
+ "loss": 4.083431243896484,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.010611358098685741,
+ "learning_rate": 0.0013938379418212863,
+ "loss": 4.042330265045166,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.009132112376391888,
+ "learning_rate": 0.001392509502019852,
+ "loss": 4.058825969696045,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.009808709844946861,
+ "learning_rate": 0.0013911811609067992,
+ "loss": 4.093010902404785,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.4473161697387695,
+ "eval_runtime": 51.7738,
+ "eval_samples_per_second": 47.167,
+ "eval_steps_per_second": 1.487,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.011859052814543247,
+ "learning_rate": 0.0013898529195523488,
+ "loss": 3.9572651386260986,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.013210238888859749,
+ "learning_rate": 0.0013885247790266406,
+ "loss": 3.9812350273132324,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.012981324456632137,
+ "learning_rate": 0.001387196740399733,
+ "loss": 3.987725019454956,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.01424382347613573,
+ "learning_rate": 0.001385868804741604,
+ "loss": 3.9719889163970947,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.0154642965644598,
+ "learning_rate": 0.0013845409731221465,
+ "loss": 3.984248161315918,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.014779271557927132,
+ "learning_rate": 0.0013832132466111706,
+ "loss": 4.002912521362305,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.012921188026666641,
+ "learning_rate": 0.0013818856262784013,
+ "loss": 3.9588499069213867,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01165325753390789,
+ "learning_rate": 0.0013805581131934792,
+ "loss": 3.993288993835449,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.008908954448997974,
+ "learning_rate": 0.0013792307084259563,
+ "loss": 3.987151861190796,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.012040507979691029,
+ "learning_rate": 0.0013779034130452991,
+ "loss": 3.9675703048706055,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.013464349322021008,
+ "learning_rate": 0.0013765762281208868,
+ "loss": 3.9901437759399414,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.009626250714063644,
+ "learning_rate": 0.0013752491547220063,
+ "loss": 3.9373281002044678,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.00897257961332798,
+ "learning_rate": 0.0013739221939178582,
+ "loss": 4.00804328918457,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.010689414106309414,
+ "learning_rate": 0.00137259534677755,
+ "loss": 4.020804405212402,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.010462156496942043,
+ "learning_rate": 0.0013712686143700988,
+ "loss": 4.028973579406738,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.009135857224464417,
+ "learning_rate": 0.001369941997764429,
+ "loss": 3.9798216819763184,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.009298899210989475,
+ "learning_rate": 0.0013686154980293713,
+ "loss": 4.018845081329346,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.00979928020387888,
+ "learning_rate": 0.0013672891162336636,
+ "loss": 3.9946775436401367,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.008175542578101158,
+ "learning_rate": 0.0013659628534459457,
+ "loss": 3.981858491897583,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.008302642032504082,
+ "learning_rate": 0.0013646367107347662,
+ "loss": 3.991128444671631,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.008516009896993637,
+ "learning_rate": 0.0013633106891685718,
+ "loss": 4.004990100860596,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.007936482317745686,
+ "learning_rate": 0.0013619847898157164,
+ "loss": 3.995734214782715,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.008317600004374981,
+ "learning_rate": 0.001360659013744451,
+ "loss": 4.040653228759766,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.008481282740831375,
+ "learning_rate": 0.001359333362022931,
+ "loss": 3.969860553741455,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.008944948203861713,
+ "learning_rate": 0.0013580078357192096,
+ "loss": 3.9908909797668457,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.008897850289940834,
+ "learning_rate": 0.001356682435901239,
+ "loss": 4.04786491394043,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.009527456946671009,
+ "learning_rate": 0.001355357163636871,
+ "loss": 4.017423629760742,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.011714582331478596,
+ "learning_rate": 0.001354032019993852,
+ "loss": 3.9990692138671875,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.011083807796239853,
+ "learning_rate": 0.0013527070060398274,
+ "loss": 3.995265483856201,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.009878288954496384,
+ "learning_rate": 0.0013513821228423363,
+ "loss": 4.0364532470703125,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.009727329947054386,
+ "learning_rate": 0.0013500573714688136,
+ "loss": 4.035953521728516,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.009482033550739288,
+ "learning_rate": 0.0013487327529865879,
+ "loss": 4.015855312347412,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.009037062525749207,
+ "learning_rate": 0.0013474082684628798,
+ "loss": 4.024164199829102,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.009094092063605785,
+ "learning_rate": 0.0013460839189648028,
+ "loss": 4.037765979766846,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.008230654522776604,
+ "learning_rate": 0.0013447597055593612,
+ "loss": 4.006317615509033,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.008491289801895618,
+ "learning_rate": 0.0013434356293134501,
+ "loss": 4.042603969573975,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.008907693438231945,
+ "learning_rate": 0.0013421116912938533,
+ "loss": 4.008098602294922,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.008897818624973297,
+ "learning_rate": 0.0013407878925672438,
+ "loss": 3.9892945289611816,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.00957555789500475,
+ "learning_rate": 0.0013394642342001832,
+ "loss": 4.046009063720703,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.009948568418622017,
+ "learning_rate": 0.0013381407172591185,
+ "loss": 3.9760642051696777,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.01014955248683691,
+ "learning_rate": 0.0013368173428103835,
+ "loss": 4.023215293884277,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.008578895591199398,
+ "learning_rate": 0.0013354941119201966,
+ "loss": 4.002819061279297,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.008729875087738037,
+ "learning_rate": 0.001334171025654662,
+ "loss": 4.0340657234191895,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.009556280449032784,
+ "learning_rate": 0.0013328480850797655,
+ "loss": 3.983076333999634,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.009733820334076881,
+ "learning_rate": 0.0013315252912613766,
+ "loss": 4.066578388214111,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.009025252424180508,
+ "learning_rate": 0.0013302026452652474,
+ "loss": 3.9980697631835938,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.008485320024192333,
+ "learning_rate": 0.001328880148157008,
+ "loss": 4.018826961517334,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.009539696387946606,
+ "learning_rate": 0.0013275578010021726,
+ "loss": 4.011737823486328,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.010881016030907631,
+ "learning_rate": 0.0013262356048661305,
+ "loss": 4.007291793823242,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.010362236760556698,
+ "learning_rate": 0.0013249135608141523,
+ "loss": 3.9929394721984863,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.010296608321368694,
+ "learning_rate": 0.0013235916699113849,
+ "loss": 4.0015130043029785,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.011211174540221691,
+ "learning_rate": 0.0013222699332228517,
+ "loss": 4.005324363708496,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.009743994101881981,
+ "learning_rate": 0.0013209483518134525,
+ "loss": 4.0278143882751465,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.008563914336264133,
+ "learning_rate": 0.0013196269267479607,
+ "loss": 3.9997801780700684,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.009078902192413807,
+ "learning_rate": 0.0013183056590910255,
+ "loss": 4.026200294494629,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.009618895128369331,
+ "learning_rate": 0.0013169845499071675,
+ "loss": 4.027050018310547,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.008920644409954548,
+ "learning_rate": 0.0013156636002607822,
+ "loss": 4.000490665435791,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.008849202655255795,
+ "learning_rate": 0.0013143428112161323,
+ "loss": 4.044985294342041,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.009303621016442776,
+ "learning_rate": 0.001313022183837356,
+ "loss": 4.046281814575195,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.01076227892190218,
+ "learning_rate": 0.0013117017191884579,
+ "loss": 4.030576705932617,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.010400841943919659,
+ "learning_rate": 0.0013103814183333127,
+ "loss": 3.980081558227539,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.009459166787564754,
+ "learning_rate": 0.001309061282335663,
+ "loss": 4.014824867248535,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.009054969064891338,
+ "learning_rate": 0.001307741312259118,
+ "loss": 4.0485944747924805,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.009003991261124611,
+ "learning_rate": 0.0013064215091671545,
+ "loss": 3.9990320205688477,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.008959448896348476,
+ "learning_rate": 0.001305101874123113,
+ "loss": 4.038620948791504,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.009102243930101395,
+ "learning_rate": 0.0013037824081902015,
+ "loss": 4.030906677246094,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.008599337190389633,
+ "learning_rate": 0.0013024631124314877,
+ "loss": 4.02648401260376,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.008771977387368679,
+ "learning_rate": 0.0013011439879099056,
+ "loss": 4.027151107788086,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.009082595817744732,
+ "learning_rate": 0.00129982503568825,
+ "loss": 4.0194478034973145,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.009025354869663715,
+ "learning_rate": 0.0012985062568291763,
+ "loss": 4.044382095336914,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.00875102635473013,
+ "learning_rate": 0.0012971876523952014,
+ "loss": 4.0257568359375,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.009312704205513,
+ "learning_rate": 0.0012958692234487006,
+ "loss": 3.9964494705200195,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.431707859039307,
+ "eval_runtime": 50.7311,
+ "eval_samples_per_second": 48.136,
+ "eval_steps_per_second": 1.518,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.00992550328373909,
+ "learning_rate": 0.0012945509710519088,
+ "loss": 3.9253792762756348,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.009920453652739525,
+ "learning_rate": 0.0012932328962669173,
+ "loss": 3.9438819885253906,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.009437218308448792,
+ "learning_rate": 0.0012919150001556756,
+ "loss": 3.960963249206543,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.009616893716156483,
+ "learning_rate": 0.0012905972837799896,
+ "loss": 3.969301223754883,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.009107022546231747,
+ "learning_rate": 0.0012892797482015185,
+ "loss": 3.9626078605651855,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.011317167431116104,
+ "learning_rate": 0.0012879623944817772,
+ "loss": 3.946483850479126,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.010536770336329937,
+ "learning_rate": 0.001286645223682134,
+ "loss": 3.959566116333008,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.009349033236503601,
+ "learning_rate": 0.0012853282368638097,
+ "loss": 3.964435338973999,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.009950080886483192,
+ "learning_rate": 0.0012840114350878762,
+ "loss": 3.9729413986206055,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.010833023115992546,
+ "learning_rate": 0.001282694819415257,
+ "loss": 4.006421089172363,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.011538525111973286,
+ "learning_rate": 0.0012813783909067266,
+ "loss": 3.9558355808258057,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.010671315714716911,
+ "learning_rate": 0.0012800621506229065,
+ "loss": 4.0052337646484375,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.010835555382072926,
+ "learning_rate": 0.0012787460996242686,
+ "loss": 3.958641767501831,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.01082367543131113,
+ "learning_rate": 0.0012774302389711304,
+ "loss": 3.9818902015686035,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.009194346144795418,
+ "learning_rate": 0.0012761145697236583,
+ "loss": 3.9688854217529297,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.007859935984015465,
+ "learning_rate": 0.001274799092941862,
+ "loss": 3.9730639457702637,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.008534743450582027,
+ "learning_rate": 0.0012734838096855978,
+ "loss": 3.975003719329834,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.009369445964694023,
+ "learning_rate": 0.0012721687210145666,
+ "loss": 3.972907304763794,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.0090147340670228,
+ "learning_rate": 0.0012708538279883098,
+ "loss": 3.964256525039673,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.007745139300823212,
+ "learning_rate": 0.0012695391316662151,
+ "loss": 3.9504833221435547,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.008920101448893547,
+ "learning_rate": 0.0012682246331075075,
+ "loss": 3.9735097885131836,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.008987226523458958,
+ "learning_rate": 0.0012669103333712568,
+ "loss": 3.9482927322387695,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.00841289572417736,
+ "learning_rate": 0.0012655962335163685,
+ "loss": 3.965221881866455,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.008016116917133331,
+ "learning_rate": 0.0012642823346015902,
+ "loss": 3.9390206336975098,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.008470152504742146,
+ "learning_rate": 0.0012629686376855068,
+ "loss": 3.9257445335388184,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.008637234568595886,
+ "learning_rate": 0.0012616551438265394,
+ "loss": 3.9999356269836426,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.008036923594772816,
+ "learning_rate": 0.0012603418540829473,
+ "loss": 3.9755172729492188,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.008908843621611595,
+ "learning_rate": 0.0012590287695128232,
+ "loss": 3.9248545169830322,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.008743193000555038,
+ "learning_rate": 0.001257715891174096,
+ "loss": 3.9619390964508057,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.00834199320524931,
+ "learning_rate": 0.001256403220124528,
+ "loss": 3.961569309234619,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.009148893877863884,
+ "learning_rate": 0.001255090757421715,
+ "loss": 3.972104549407959,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.00925084762275219,
+ "learning_rate": 0.0012537785041230843,
+ "loss": 4.000775337219238,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.008659154176712036,
+ "learning_rate": 0.0012524664612858947,
+ "loss": 3.954564332962036,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.007787859998643398,
+ "learning_rate": 0.0012511546299672355,
+ "loss": 3.9498860836029053,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.00817918125540018,
+ "learning_rate": 0.0012498430112240246,
+ "loss": 4.003897666931152,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.009282548911869526,
+ "learning_rate": 0.001248531606113011,
+ "loss": 3.9621925354003906,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.009375610388815403,
+ "learning_rate": 0.001247220415690769,
+ "loss": 3.9977524280548096,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.009650378488004208,
+ "learning_rate": 0.001245909441013701,
+ "loss": 3.9882829189300537,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.009350919164717197,
+ "learning_rate": 0.0012445986831380366,
+ "loss": 4.019596099853516,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.008726303465664387,
+ "learning_rate": 0.0012432881431198289,
+ "loss": 3.9933621883392334,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.00829484686255455,
+ "learning_rate": 0.0012419778220149568,
+ "loss": 4.001237869262695,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.008744452148675919,
+ "learning_rate": 0.0012406677208791218,
+ "loss": 3.9323887825012207,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.008170492947101593,
+ "learning_rate": 0.0012393578407678496,
+ "loss": 3.925581455230713,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.007843422703444958,
+ "learning_rate": 0.001238048182736486,
+ "loss": 4.001023769378662,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.008914309553802013,
+ "learning_rate": 0.0012367387478401987,
+ "loss": 3.9804537296295166,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.009736751206219196,
+ "learning_rate": 0.0012354295371339774,
+ "loss": 3.9831767082214355,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.010061381384730339,
+ "learning_rate": 0.0012341205516726273,
+ "loss": 3.9462435245513916,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.010350692085921764,
+ "learning_rate": 0.0012328117925107765,
+ "loss": 4.012863636016846,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01039902213960886,
+ "learning_rate": 0.0012315032607028667,
+ "loss": 3.9811301231384277,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.011423371732234955,
+ "learning_rate": 0.0012301949573031595,
+ "loss": 3.9773542881011963,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.012449614703655243,
+ "learning_rate": 0.001228886883365731,
+ "loss": 3.969600200653076,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.010361429303884506,
+ "learning_rate": 0.0012275790399444726,
+ "loss": 4.035360336303711,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.009603082202374935,
+ "learning_rate": 0.0012262714280930904,
+ "loss": 3.9840846061706543,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.009183845482766628,
+ "learning_rate": 0.0012249640488651036,
+ "loss": 4.031542778015137,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.009694446809589863,
+ "learning_rate": 0.0012236569033138441,
+ "loss": 3.9935057163238525,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.0097816726192832,
+ "learning_rate": 0.0012223499924924545,
+ "loss": 3.9851176738739014,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.009413184598088264,
+ "learning_rate": 0.001221043317453891,
+ "loss": 3.9642205238342285,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.009960711002349854,
+ "learning_rate": 0.0012197368792509166,
+ "loss": 3.9373817443847656,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.010124670341610909,
+ "learning_rate": 0.0012184306789361059,
+ "loss": 3.958115577697754,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.011087656021118164,
+ "learning_rate": 0.00121712471756184,
+ "loss": 3.9607057571411133,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.009812800213694572,
+ "learning_rate": 0.001215818996180309,
+ "loss": 4.00722599029541,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.010305573232471943,
+ "learning_rate": 0.0012145135158435094,
+ "loss": 3.98728084564209,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.01019787136465311,
+ "learning_rate": 0.0012132082776032422,
+ "loss": 3.981172800064087,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.009260181337594986,
+ "learning_rate": 0.001211903282511115,
+ "loss": 4.034339904785156,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.00972998607903719,
+ "learning_rate": 0.001210598531618538,
+ "loss": 4.022885322570801,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.009853586554527283,
+ "learning_rate": 0.0012092940259767267,
+ "loss": 4.000235080718994,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.00961449183523655,
+ "learning_rate": 0.0012079897666366962,
+ "loss": 3.9951353073120117,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.010257176123559475,
+ "learning_rate": 0.0012066857546492657,
+ "loss": 3.9801371097564697,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.009443561546504498,
+ "learning_rate": 0.001205381991065054,
+ "loss": 3.9596502780914307,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.010389018803834915,
+ "learning_rate": 0.0012040784769344795,
+ "loss": 3.980527639389038,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.010322567075490952,
+ "learning_rate": 0.0012027752133077607,
+ "loss": 3.996258497238159,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.009987728670239449,
+ "learning_rate": 0.0012014722012349125,
+ "loss": 3.9493203163146973,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.421086311340332,
+ "eval_runtime": 52.5623,
+ "eval_samples_per_second": 46.459,
+ "eval_steps_per_second": 1.465,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.009405115619301796,
+ "learning_rate": 0.0012001694417657493,
+ "loss": 3.927506923675537,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.009974643588066101,
+ "learning_rate": 0.00119886693594988,
+ "loss": 3.9373550415039062,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.011331096291542053,
+ "learning_rate": 0.0011975646848367098,
+ "loss": 3.935594081878662,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.010538448579609394,
+ "learning_rate": 0.0011962626894754405,
+ "loss": 3.911579132080078,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.00989572424441576,
+ "learning_rate": 0.0011949609509150645,
+ "loss": 3.92525577545166,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.010533696971833706,
+ "learning_rate": 0.0011936594702043704,
+ "loss": 3.898075580596924,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.01051455456763506,
+ "learning_rate": 0.0011923582483919365,
+ "loss": 3.93463134765625,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.009662075899541378,
+ "learning_rate": 0.0011910572865261348,
+ "loss": 3.898221731185913,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.008874096907675266,
+ "learning_rate": 0.0011897565856551257,
+ "loss": 3.926787853240967,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.008126807399094105,
+ "learning_rate": 0.0011884561468268606,
+ "loss": 3.9158413410186768,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.007915553636848927,
+ "learning_rate": 0.001187155971089081,
+ "loss": 3.8891496658325195,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.008312922902405262,
+ "learning_rate": 0.0011858560594893129,
+ "loss": 3.932219982147217,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.008892328478395939,
+ "learning_rate": 0.0011845564130748734,
+ "loss": 3.8866336345672607,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.009187526069581509,
+ "learning_rate": 0.0011832570328928628,
+ "loss": 3.9464707374572754,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.008615494705736637,
+ "learning_rate": 0.0011819579199901693,
+ "loss": 3.9396629333496094,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.008649308234453201,
+ "learning_rate": 0.0011806590754134636,
+ "loss": 3.9325060844421387,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.008581082336604595,
+ "learning_rate": 0.0011793605002092021,
+ "loss": 3.9200944900512695,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.008502699434757233,
+ "learning_rate": 0.0011780621954236238,
+ "loss": 3.9656100273132324,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.0083576999604702,
+ "learning_rate": 0.0011767641621027477,
+ "loss": 3.9130215644836426,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.008196269161999226,
+ "learning_rate": 0.0011754664012923785,
+ "loss": 3.942369222640991,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.008857307955622673,
+ "learning_rate": 0.0011741689140380956,
+ "loss": 3.909585952758789,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.008935349993407726,
+ "learning_rate": 0.0011728717013852635,
+ "loss": 3.9248366355895996,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.008821200579404831,
+ "learning_rate": 0.0011715747643790211,
+ "loss": 3.9589288234710693,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.00861942209303379,
+ "learning_rate": 0.0011702781040642885,
+ "loss": 3.9381728172302246,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.009127574972808361,
+ "learning_rate": 0.0011689817214857613,
+ "loss": 3.897829532623291,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.008008269593119621,
+ "learning_rate": 0.0011676856176879107,
+ "loss": 3.930056095123291,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.010070208460092545,
+ "learning_rate": 0.0011663897937149853,
+ "loss": 3.892270803451538,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.00903113093227148,
+ "learning_rate": 0.001165094250611006,
+ "loss": 3.9766440391540527,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.008872415870428085,
+ "learning_rate": 0.0011637989894197692,
+ "loss": 3.9559712409973145,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.009524816647171974,
+ "learning_rate": 0.0011625040111848427,
+ "loss": 3.9607882499694824,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.009836779907345772,
+ "learning_rate": 0.0011612093169495678,
+ "loss": 3.9629592895507812,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.009586540050804615,
+ "learning_rate": 0.0011599149077570561,
+ "loss": 3.95208740234375,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.010495901107788086,
+ "learning_rate": 0.0011586207846501896,
+ "loss": 3.9439730644226074,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.010346266441047192,
+ "learning_rate": 0.0011573269486716202,
+ "loss": 3.959299087524414,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.010767634026706219,
+ "learning_rate": 0.0011560334008637674,
+ "loss": 3.900784969329834,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.012122200801968575,
+ "learning_rate": 0.0011547401422688202,
+ "loss": 3.9588193893432617,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.012914645485579967,
+ "learning_rate": 0.0011534471739287331,
+ "loss": 3.9278407096862793,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.011785333044826984,
+ "learning_rate": 0.0011521544968852273,
+ "loss": 3.992793560028076,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.008455781266093254,
+ "learning_rate": 0.0011508621121797899,
+ "loss": 3.96525239944458,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.009655282832682133,
+ "learning_rate": 0.0011495700208536714,
+ "loss": 3.9346890449523926,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.010537004098296165,
+ "learning_rate": 0.0011482782239478871,
+ "loss": 3.9534668922424316,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.009438651613891125,
+ "learning_rate": 0.0011469867225032135,
+ "loss": 3.983487606048584,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.010586646385490894,
+ "learning_rate": 0.0011456955175601908,
+ "loss": 3.939680576324463,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.009082718752324581,
+ "learning_rate": 0.001144404610159119,
+ "loss": 3.9558515548706055,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.009833602234721184,
+ "learning_rate": 0.0011431140013400591,
+ "loss": 3.9539828300476074,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.009142170660197735,
+ "learning_rate": 0.0011418236921428313,
+ "loss": 3.9442849159240723,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.008603853173553944,
+ "learning_rate": 0.0011405336836070137,
+ "loss": 3.9526913166046143,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.008838302455842495,
+ "learning_rate": 0.0011392439767719446,
+ "loss": 3.973104953765869,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.008867006748914719,
+ "learning_rate": 0.0011379545726767156,
+ "loss": 3.9832305908203125,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.00971697736531496,
+ "learning_rate": 0.001136665472360178,
+ "loss": 3.9563937187194824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.00957175437361002,
+ "learning_rate": 0.001135376676860935,
+ "loss": 3.975194215774536,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.008548660203814507,
+ "learning_rate": 0.001134088187217347,
+ "loss": 3.9839985370635986,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.008556334301829338,
+ "learning_rate": 0.0011328000044675264,
+ "loss": 3.9225516319274902,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.008440867997705936,
+ "learning_rate": 0.0011315121296493385,
+ "loss": 3.945028066635132,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.009643333032727242,
+ "learning_rate": 0.0011302245638004017,
+ "loss": 3.943120002746582,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.008832835592329502,
+ "learning_rate": 0.001128937307958083,
+ "loss": 3.9895920753479004,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.008720474317669868,
+ "learning_rate": 0.0011276503631595036,
+ "loss": 3.941950798034668,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.008099926635622978,
+ "learning_rate": 0.001126363730441529,
+ "loss": 3.9626870155334473,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.008525107055902481,
+ "learning_rate": 0.0011250774108407778,
+ "loss": 3.950535774230957,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.008998272009193897,
+ "learning_rate": 0.0011237914053936135,
+ "loss": 3.9460678100585938,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.009387437254190445,
+ "learning_rate": 0.001122505715136148,
+ "loss": 3.926604986190796,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.007804017513990402,
+ "learning_rate": 0.0011212203411042386,
+ "loss": 3.9894230365753174,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.008564616553485394,
+ "learning_rate": 0.0011199352843334876,
+ "loss": 3.9339959621429443,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.008726440370082855,
+ "learning_rate": 0.0011186505458592427,
+ "loss": 3.9867987632751465,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.008074100129306316,
+ "learning_rate": 0.0011173661267165932,
+ "loss": 3.9492292404174805,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.007688041776418686,
+ "learning_rate": 0.0011160820279403744,
+ "loss": 3.9425454139709473,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.007808993570506573,
+ "learning_rate": 0.0011147982505651595,
+ "loss": 3.935114860534668,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.009786820039153099,
+ "learning_rate": 0.0011135147956252659,
+ "loss": 3.959758758544922,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.009487155824899673,
+ "learning_rate": 0.00111223166415475,
+ "loss": 3.989048480987549,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.008407804183661938,
+ "learning_rate": 0.0011109488571874073,
+ "loss": 3.932347297668457,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.007158380467444658,
+ "learning_rate": 0.0011096663757567723,
+ "loss": 3.9545154571533203,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.008840581402182579,
+ "learning_rate": 0.001108384220896117,
+ "loss": 3.965984344482422,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.409194469451904,
+ "eval_runtime": 51.5561,
+ "eval_samples_per_second": 47.366,
+ "eval_steps_per_second": 1.494,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.010702896863222122,
+ "learning_rate": 0.0011071023936384508,
+ "loss": 3.8964462280273438,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01290619745850563,
+ "learning_rate": 0.0011058208950165183,
+ "loss": 3.8775625228881836,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01443667896091938,
+ "learning_rate": 0.0011045397260628,
+ "loss": 3.8755288124084473,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.014411968179047108,
+ "learning_rate": 0.0011032588878095108,
+ "loss": 3.889918327331543,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.013656266033649445,
+ "learning_rate": 0.0011019783812885986,
+ "loss": 3.926997661590576,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.010661937296390533,
+ "learning_rate": 0.0011006982075317446,
+ "loss": 3.9226956367492676,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.008968768641352654,
+ "learning_rate": 0.0010994183675703615,
+ "loss": 3.8932647705078125,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.009992181323468685,
+ "learning_rate": 0.0010981388624355937,
+ "loss": 3.8920068740844727,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.011034460738301277,
+ "learning_rate": 0.0010968596931583148,
+ "loss": 3.909073829650879,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.011815821751952171,
+ "learning_rate": 0.0010955808607691284,
+ "loss": 3.9190006256103516,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.012317490763962269,
+ "learning_rate": 0.0010943023662983674,
+ "loss": 3.8790552616119385,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.010983958840370178,
+ "learning_rate": 0.0010930242107760903,
+ "loss": 3.8730852603912354,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.008948802947998047,
+ "learning_rate": 0.0010917463952320856,
+ "loss": 3.923659563064575,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.009568693116307259,
+ "learning_rate": 0.0010904689206958649,
+ "loss": 3.9100794792175293,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.011862963438034058,
+ "learning_rate": 0.0010891917881966672,
+ "loss": 3.908093214035034,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.010180995799601078,
+ "learning_rate": 0.0010879149987634545,
+ "loss": 3.900660514831543,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.009372464381158352,
+ "learning_rate": 0.0010866385534249131,
+ "loss": 3.8950703144073486,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.010695629753172398,
+ "learning_rate": 0.0010853624532094525,
+ "loss": 3.913530111312866,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.011817459017038345,
+ "learning_rate": 0.0010840866991452024,
+ "loss": 3.90268611907959,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.010686337016522884,
+ "learning_rate": 0.0010828112922600167,
+ "loss": 3.9141364097595215,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.009550466202199459,
+ "learning_rate": 0.0010815362335814655,
+ "loss": 3.968468189239502,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.009367208927869797,
+ "learning_rate": 0.0010802615241368423,
+ "loss": 3.8933968544006348,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.009899397380650043,
+ "learning_rate": 0.0010789871649531562,
+ "loss": 3.936708450317383,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.009181005880236626,
+ "learning_rate": 0.0010777131570571363,
+ "loss": 3.9370503425598145,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.00799303874373436,
+ "learning_rate": 0.0010764395014752274,
+ "loss": 3.911050319671631,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.008469315245747566,
+ "learning_rate": 0.0010751661992335908,
+ "loss": 3.9103240966796875,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.009814459830522537,
+ "learning_rate": 0.0010738932513581034,
+ "loss": 3.881843090057373,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.010244108736515045,
+ "learning_rate": 0.0010726206588743557,
+ "loss": 3.883141040802002,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.008898732252418995,
+ "learning_rate": 0.001071348422807653,
+ "loss": 3.952347755432129,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.00823870487511158,
+ "learning_rate": 0.0010700765441830121,
+ "loss": 3.932507038116455,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.009857657365500927,
+ "learning_rate": 0.001068805024025164,
+ "loss": 3.932932138442993,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.00929308496415615,
+ "learning_rate": 0.0010675338633585488,
+ "loss": 3.8971338272094727,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.008667961694300175,
+ "learning_rate": 0.0010662630632073172,
+ "loss": 3.916931629180908,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.009758287109434605,
+ "learning_rate": 0.001064992624595331,
+ "loss": 3.964726209640503,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.009429254569113255,
+ "learning_rate": 0.0010637225485461586,
+ "loss": 3.917572498321533,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.008303353562951088,
+ "learning_rate": 0.0010624528360830782,
+ "loss": 3.875241756439209,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.008729013614356518,
+ "learning_rate": 0.001061183488229073,
+ "loss": 3.9691600799560547,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.010239302180707455,
+ "learning_rate": 0.0010599145060068353,
+ "loss": 3.9053149223327637,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.00932721421122551,
+ "learning_rate": 0.0010586458904387596,
+ "loss": 3.8593955039978027,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.008196988143026829,
+ "learning_rate": 0.0010573776425469473,
+ "loss": 3.917983055114746,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.008049141615629196,
+ "learning_rate": 0.001056109763353203,
+ "loss": 3.9337785243988037,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.009256313554942608,
+ "learning_rate": 0.001054842253879034,
+ "loss": 3.937887668609619,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.009292179718613625,
+ "learning_rate": 0.0010535751151456495,
+ "loss": 3.9203450679779053,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.009142161346971989,
+ "learning_rate": 0.0010523083481739602,
+ "loss": 3.9027457237243652,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.009498288854956627,
+ "learning_rate": 0.0010510419539845777,
+ "loss": 3.917316436767578,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.00849096104502678,
+ "learning_rate": 0.001049775933597813,
+ "loss": 3.9229235649108887,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.008343606255948544,
+ "learning_rate": 0.0010485102880336747,
+ "loss": 3.9138073921203613,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.010342134162783623,
+ "learning_rate": 0.0010472450183118727,
+ "loss": 3.8932929039001465,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.010298306122422218,
+ "learning_rate": 0.00104598012545181,
+ "loss": 3.9112677574157715,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.00827560480684042,
+ "learning_rate": 0.001044715610472589,
+ "loss": 3.894556760787964,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.008993102237582207,
+ "learning_rate": 0.0010434514743930064,
+ "loss": 3.905391216278076,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.008632538840174675,
+ "learning_rate": 0.0010421877182315536,
+ "loss": 3.896714210510254,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.008542424067854881,
+ "learning_rate": 0.0010409243430064162,
+ "loss": 3.9145569801330566,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.008970450609922409,
+ "learning_rate": 0.0010396613497354725,
+ "loss": 3.9170360565185547,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.008744142949581146,
+ "learning_rate": 0.0010383987394362942,
+ "loss": 3.927412509918213,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.008229847066104412,
+ "learning_rate": 0.0010371365131261424,
+ "loss": 3.928352117538452,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.00818693358451128,
+ "learning_rate": 0.0010358746718219717,
+ "loss": 3.9219343662261963,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.008848588913679123,
+ "learning_rate": 0.001034613216540423,
+ "loss": 3.9346699714660645,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.00834780652076006,
+ "learning_rate": 0.0010333521482978295,
+ "loss": 3.9144492149353027,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.00940527394413948,
+ "learning_rate": 0.0010320914681102098,
+ "loss": 3.951573371887207,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.008884521201252937,
+ "learning_rate": 0.0010308311769932725,
+ "loss": 3.899625778198242,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.007863261736929417,
+ "learning_rate": 0.0010295712759624107,
+ "loss": 3.9321727752685547,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.008651153184473515,
+ "learning_rate": 0.0010283117660327035,
+ "loss": 3.9030094146728516,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.00822074618190527,
+ "learning_rate": 0.0010270526482189163,
+ "loss": 3.939466953277588,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.008701153099536896,
+ "learning_rate": 0.0010257939235354964,
+ "loss": 3.9099626541137695,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.008418620564043522,
+ "learning_rate": 0.0010245355929965774,
+ "loss": 3.914116382598877,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.008339700289070606,
+ "learning_rate": 0.001023277657615971,
+ "loss": 3.9111878871917725,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.008304647170007229,
+ "learning_rate": 0.0010220201184071747,
+ "loss": 3.950974464416504,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.008436229079961777,
+ "learning_rate": 0.001020762976383365,
+ "loss": 3.9631457328796387,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.0074487063102424145,
+ "learning_rate": 0.001019506232557398,
+ "loss": 3.902984619140625,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.008736515417695045,
+ "learning_rate": 0.0010182498879418096,
+ "loss": 3.9268949031829834,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.009666468016803265,
+ "learning_rate": 0.0010169939435488145,
+ "loss": 3.9274301528930664,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.395970821380615,
+ "eval_runtime": 53.7848,
+ "eval_samples_per_second": 45.403,
+ "eval_steps_per_second": 1.432,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.009094545617699623,
+ "learning_rate": 0.001015738400390304,
+ "loss": 3.866425037384033,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.009273733012378216,
+ "learning_rate": 0.0010144832594778462,
+ "loss": 3.8569841384887695,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.009205271489918232,
+ "learning_rate": 0.0010132285218226855,
+ "loss": 3.8981378078460693,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.009963045828044415,
+ "learning_rate": 0.0010119741884357424,
+ "loss": 3.881542444229126,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.008536554872989655,
+ "learning_rate": 0.0010107202603276096,
+ "loss": 3.851253032684326,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.01074561569839716,
+ "learning_rate": 0.001009466738508555,
+ "loss": 3.8731184005737305,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.010809902101755142,
+ "learning_rate": 0.0010082136239885177,
+ "loss": 3.863889217376709,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.008289352990686893,
+ "learning_rate": 0.0010069609177771103,
+ "loss": 3.839066743850708,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.00792208593338728,
+ "learning_rate": 0.001005708620883615,
+ "loss": 3.8634843826293945,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.008381598629057407,
+ "learning_rate": 0.0010044567343169845,
+ "loss": 3.868283987045288,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.009784617461264133,
+ "learning_rate": 0.0010032052590858425,
+ "loss": 3.853074550628662,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.009462198242545128,
+ "learning_rate": 0.0010019541961984782,
+ "loss": 3.889979600906372,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.010207819752395153,
+ "learning_rate": 0.001000703546662852,
+ "loss": 3.8636741638183594,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.009123551659286022,
+ "learning_rate": 0.0009994533114865887,
+ "loss": 3.8578555583953857,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.009006140753626823,
+ "learning_rate": 0.0009982034916769806,
+ "loss": 3.8828015327453613,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.0091114342212677,
+ "learning_rate": 0.0009969540882409846,
+ "loss": 3.880486011505127,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.008851367980241776,
+ "learning_rate": 0.0009957051021852227,
+ "loss": 3.8618125915527344,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.008196741342544556,
+ "learning_rate": 0.0009944565345159806,
+ "loss": 3.8314175605773926,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.00863513071089983,
+ "learning_rate": 0.000993208386239206,
+ "loss": 3.856212615966797,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.008624681271612644,
+ "learning_rate": 0.0009919606583605107,
+ "loss": 3.868927001953125,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.009208842180669308,
+ "learning_rate": 0.0009907133518851647,
+ "loss": 3.9072189331054688,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.010200913064181805,
+ "learning_rate": 0.0009894664678181027,
+ "loss": 3.875272750854492,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.009892710484564304,
+ "learning_rate": 0.000988220007163914,
+ "loss": 3.888686418533325,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.009873585775494576,
+ "learning_rate": 0.000986973970926851,
+ "loss": 3.8607118129730225,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.009322221390902996,
+ "learning_rate": 0.0009857283601108228,
+ "loss": 3.8852052688598633,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.008141528815031052,
+ "learning_rate": 0.0009844831757193947,
+ "loss": 3.888213872909546,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.008191613480448723,
+ "learning_rate": 0.00098323841875579,
+ "loss": 3.9073967933654785,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.008419081568717957,
+ "learning_rate": 0.0009819940902228864,
+ "loss": 3.879223346710205,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.009037869982421398,
+ "learning_rate": 0.0009807501911232171,
+ "loss": 3.883303165435791,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.00854078121483326,
+ "learning_rate": 0.0009795067224589687,
+ "loss": 3.926767587661743,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.008590771816670895,
+ "learning_rate": 0.0009782636852319825,
+ "loss": 3.913999557495117,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.00859815627336502,
+ "learning_rate": 0.0009770210804437498,
+ "loss": 3.8828063011169434,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.008970260620117188,
+ "learning_rate": 0.0009757789090954157,
+ "loss": 3.9174866676330566,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.00877175759524107,
+ "learning_rate": 0.0009745371721877754,
+ "loss": 3.8879401683807373,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.008541764691472054,
+ "learning_rate": 0.0009732958707212731,
+ "loss": 3.8717427253723145,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.008466241881251335,
+ "learning_rate": 0.0009720550056960039,
+ "loss": 3.8944485187530518,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.00808491837233305,
+ "learning_rate": 0.0009708145781117096,
+ "loss": 3.8617026805877686,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.00842207670211792,
+ "learning_rate": 0.0009695745889677805,
+ "loss": 3.887824535369873,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.00830599945038557,
+ "learning_rate": 0.0009683350392632533,
+ "loss": 3.8612070083618164,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.008417914621531963,
+ "learning_rate": 0.0009670959299968111,
+ "loss": 3.880140781402588,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.008542059920728207,
+ "learning_rate": 0.0009658572621667822,
+ "loss": 3.8951029777526855,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.008822236210107803,
+ "learning_rate": 0.000964619036771138,
+ "loss": 3.8962583541870117,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.009151184931397438,
+ "learning_rate": 0.0009633812548074952,
+ "loss": 3.882500648498535,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.007803036365658045,
+ "learning_rate": 0.0009621439172731115,
+ "loss": 3.8740782737731934,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.009350850246846676,
+ "learning_rate": 0.0009609070251648884,
+ "loss": 3.8627419471740723,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.008319144137203693,
+ "learning_rate": 0.0009596705794793667,
+ "loss": 3.9084486961364746,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.009198528714478016,
+ "learning_rate": 0.0009584345812127282,
+ "loss": 3.92141056060791,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.010480938479304314,
+ "learning_rate": 0.0009571990313607958,
+ "loss": 3.8916103839874268,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.01045288611203432,
+ "learning_rate": 0.0009559639309190278,
+ "loss": 3.8863210678100586,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.009121391922235489,
+ "learning_rate": 0.0009547292808825239,
+ "loss": 3.9167044162750244,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.01093223039060831,
+ "learning_rate": 0.0009534950822460184,
+ "loss": 3.9046711921691895,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.01214868575334549,
+ "learning_rate": 0.0009522613360038835,
+ "loss": 3.9126768112182617,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.01055280864238739,
+ "learning_rate": 0.0009510280431501255,
+ "loss": 3.9214353561401367,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.009021313861012459,
+ "learning_rate": 0.0009497952046783867,
+ "loss": 3.8795580863952637,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.00863752793520689,
+ "learning_rate": 0.0009485628215819428,
+ "loss": 3.893232822418213,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.009327626787126064,
+ "learning_rate": 0.0009473308948537021,
+ "loss": 3.917377471923828,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.01003898587077856,
+ "learning_rate": 0.0009460994254862066,
+ "loss": 3.8818931579589844,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.00892532430589199,
+ "learning_rate": 0.0009448684144716278,
+ "loss": 3.8761978149414062,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.008396463468670845,
+ "learning_rate": 0.0009436378628017702,
+ "loss": 3.8665928840637207,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.00784370768815279,
+ "learning_rate": 0.0009424077714680662,
+ "loss": 3.8850138187408447,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.00809321366250515,
+ "learning_rate": 0.0009411781414615784,
+ "loss": 3.917078733444214,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.008685288950800896,
+ "learning_rate": 0.000939948973772998,
+ "loss": 3.8865013122558594,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.007291820365935564,
+ "learning_rate": 0.0009387202693926425,
+ "loss": 3.8967666625976562,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.008459319360554218,
+ "learning_rate": 0.0009374920293104573,
+ "loss": 3.9043235778808594,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.007924137637019157,
+ "learning_rate": 0.0009362642545160123,
+ "loss": 3.9216856956481934,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.007874754257500172,
+ "learning_rate": 0.0009350369459985052,
+ "loss": 3.879725933074951,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.008643723092973232,
+ "learning_rate": 0.0009338101047467544,
+ "loss": 3.9303886890411377,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.008451704867184162,
+ "learning_rate": 0.000932583731749205,
+ "loss": 3.9352827072143555,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.008137754164636135,
+ "learning_rate": 0.0009313578279939236,
+ "loss": 3.930800437927246,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.009277567267417908,
+ "learning_rate": 0.0009301323944685979,
+ "loss": 3.9194300174713135,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.009428970515727997,
+ "learning_rate": 0.0009289074321605383,
+ "loss": 3.874298572540283,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.0089727733284235,
+ "learning_rate": 0.0009276829420566743,
+ "loss": 3.893770694732666,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.388528823852539,
+ "eval_runtime": 51.5859,
+ "eval_samples_per_second": 47.338,
+ "eval_steps_per_second": 1.493,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.009470812045037746,
+ "learning_rate": 0.000926458925143556,
+ "loss": 3.8514318466186523,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.011485518887639046,
+ "learning_rate": 0.0009252353824073511,
+ "loss": 3.857734203338623,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.010095939971506596,
+ "learning_rate": 0.0009240123148338457,
+ "loss": 3.819117546081543,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.010103728622198105,
+ "learning_rate": 0.0009227897234084444,
+ "loss": 3.842660427093506,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.01148147787898779,
+ "learning_rate": 0.0009215676091161661,
+ "loss": 3.821089029312134,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.012331553734838963,
+ "learning_rate": 0.0009203459729416467,
+ "loss": 3.8194339275360107,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.011274794116616249,
+ "learning_rate": 0.0009191248158691358,
+ "loss": 3.8839240074157715,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.009603365324437618,
+ "learning_rate": 0.0009179041388824986,
+ "loss": 3.806617259979248,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.010824795812368393,
+ "learning_rate": 0.0009166839429652112,
+ "loss": 3.823228597640991,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.01014280878007412,
+ "learning_rate": 0.0009154642291003637,
+ "loss": 3.860605239868164,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.009052124805748463,
+ "learning_rate": 0.0009142449982706591,
+ "loss": 3.84226131439209,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.009465130977332592,
+ "learning_rate": 0.0009130262514584074,
+ "loss": 3.8674912452697754,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.009991783648729324,
+ "learning_rate": 0.0009118079896455326,
+ "loss": 3.8775739669799805,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.009288253262639046,
+ "learning_rate": 0.0009105902138135652,
+ "loss": 3.7978098392486572,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.009964323602616787,
+ "learning_rate": 0.0009093729249436459,
+ "loss": 3.8362321853637695,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.010977157391607761,
+ "learning_rate": 0.0009081561240165218,
+ "loss": 3.857304096221924,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.009241319261491299,
+ "learning_rate": 0.0009069398120125479,
+ "loss": 3.8659310340881348,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.009211038239300251,
+ "learning_rate": 0.0009057239899116844,
+ "loss": 3.8764538764953613,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.009649050422012806,
+ "learning_rate": 0.000904508658693497,
+ "loss": 3.8702778816223145,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.00954380165785551,
+ "learning_rate": 0.0009032938193371573,
+ "loss": 3.8273327350616455,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.008831270970404148,
+ "learning_rate": 0.0009020794728214378,
+ "loss": 3.821664810180664,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.008367286063730717,
+ "learning_rate": 0.000900865620124717,
+ "loss": 3.8586034774780273,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.008805174380540848,
+ "learning_rate": 0.0008996522622249732,
+ "loss": 3.811522960662842,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.00873226672410965,
+ "learning_rate": 0.0008984394000997873,
+ "loss": 3.8538150787353516,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.008930203504860401,
+ "learning_rate": 0.0008972270347263402,
+ "loss": 3.8448486328125,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.009300434030592442,
+ "learning_rate": 0.0008960151670814125,
+ "loss": 3.842113494873047,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.009598663076758385,
+ "learning_rate": 0.0008948037981413855,
+ "loss": 3.886852741241455,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.008400370366871357,
+ "learning_rate": 0.0008935929288822357,
+ "loss": 3.852076292037964,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.00891344528645277,
+ "learning_rate": 0.0008923825602795391,
+ "loss": 3.8480606079101562,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.00869396049529314,
+ "learning_rate": 0.0008911726933084678,
+ "loss": 3.8164706230163574,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.007914266549050808,
+ "learning_rate": 0.0008899633289437904,
+ "loss": 3.8768272399902344,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.008398533798754215,
+ "learning_rate": 0.0008887544681598687,
+ "loss": 3.8625502586364746,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.00887146033346653,
+ "learning_rate": 0.0008875461119306601,
+ "loss": 3.8934593200683594,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.008642833679914474,
+ "learning_rate": 0.0008863382612297172,
+ "loss": 3.8391976356506348,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.008099788799881935,
+ "learning_rate": 0.0008851309170301816,
+ "loss": 3.827965497970581,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.008166342042386532,
+ "learning_rate": 0.0008839240803047894,
+ "loss": 3.8999152183532715,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.009095479734241962,
+ "learning_rate": 0.0008827177520258671,
+ "loss": 3.876927614212036,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.009246554225683212,
+ "learning_rate": 0.0008815119331653321,
+ "loss": 3.8692450523376465,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.007059223018586636,
+ "learning_rate": 0.0008803066246946898,
+ "loss": 3.869400978088379,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.008499744348227978,
+ "learning_rate": 0.0008791018275850363,
+ "loss": 3.8303189277648926,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.007667839527130127,
+ "learning_rate": 0.0008778975428070543,
+ "loss": 3.827998638153076,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.00765089550986886,
+ "learning_rate": 0.0008766937713310145,
+ "loss": 3.8554482460021973,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.00862938817590475,
+ "learning_rate": 0.0008754905141267749,
+ "loss": 3.905458927154541,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.008733712136745453,
+ "learning_rate": 0.0008742877721637758,
+ "loss": 3.8564133644104004,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.0071738120168447495,
+ "learning_rate": 0.0008730855464110471,
+ "loss": 3.878807544708252,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.008418241515755653,
+ "learning_rate": 0.0008718838378371989,
+ "loss": 3.8399720191955566,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.008526502177119255,
+ "learning_rate": 0.0008706826474104263,
+ "loss": 3.847656488418579,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.00732462527230382,
+ "learning_rate": 0.0008694819760985069,
+ "loss": 3.8839492797851562,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.008424196392297745,
+ "learning_rate": 0.0008682818248688003,
+ "loss": 3.861239433288574,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.008984955959022045,
+ "learning_rate": 0.0008670821946882465,
+ "loss": 3.856529712677002,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.009004034101963043,
+ "learning_rate": 0.000865883086523365,
+ "loss": 3.8514528274536133,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.009097260423004627,
+ "learning_rate": 0.0008646845013402563,
+ "loss": 3.8890581130981445,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.009244748391211033,
+ "learning_rate": 0.0008634864401045988,
+ "loss": 3.872745990753174,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.00788815412670374,
+ "learning_rate": 0.0008622889037816488,
+ "loss": 3.8777575492858887,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.008943610824644566,
+ "learning_rate": 0.0008610918933362396,
+ "loss": 3.8825736045837402,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.008154108189046383,
+ "learning_rate": 0.0008598954097327809,
+ "loss": 3.877514362335205,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.008689315989613533,
+ "learning_rate": 0.0008586994539352587,
+ "loss": 3.854720115661621,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.00904505793005228,
+ "learning_rate": 0.0008575040269072315,
+ "loss": 3.911463499069214,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.007773210760205984,
+ "learning_rate": 0.0008563091296118342,
+ "loss": 3.875370740890503,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.008437697775661945,
+ "learning_rate": 0.0008551147630117737,
+ "loss": 3.8457469940185547,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.009780957363545895,
+ "learning_rate": 0.0008539209280693296,
+ "loss": 3.890479564666748,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.007929213345050812,
+ "learning_rate": 0.0008527276257463532,
+ "loss": 3.876030445098877,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.009121386334300041,
+ "learning_rate": 0.0008515348570042669,
+ "loss": 3.9252734184265137,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.00981531199067831,
+ "learning_rate": 0.000850342622804063,
+ "loss": 3.8997983932495117,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.009150844067335129,
+ "learning_rate": 0.0008491509241063025,
+ "loss": 3.85115385055542,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.00992978923022747,
+ "learning_rate": 0.0008479597618711154,
+ "loss": 3.9138102531433105,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.010906689800322056,
+ "learning_rate": 0.0008467691370582001,
+ "loss": 3.892329692840576,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.008644519373774529,
+ "learning_rate": 0.0008455790506268213,
+ "loss": 3.9071311950683594,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.009030954912304878,
+ "learning_rate": 0.0008443895035358111,
+ "loss": 3.867384910583496,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.010806960985064507,
+ "learning_rate": 0.0008432004967435638,
+ "loss": 3.877983331680298,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.011148087680339813,
+ "learning_rate": 0.0008420120312080436,
+ "loss": 3.8795924186706543,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01046178862452507,
+ "learning_rate": 0.0008408241078867731,
+ "loss": 3.887709617614746,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.383835792541504,
+ "eval_runtime": 53.0778,
+ "eval_samples_per_second": 46.008,
+ "eval_steps_per_second": 1.451,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.009721867740154266,
+ "learning_rate": 0.0008396367277368421,
+ "loss": 3.8261327743530273,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.01091420091688633,
+ "learning_rate": 0.000838449891714901,
+ "loss": 3.7837276458740234,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.010423852130770683,
+ "learning_rate": 0.0008372636007771623,
+ "loss": 3.796450138092041,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.011928882449865341,
+ "learning_rate": 0.0008360778558793996,
+ "loss": 3.794717311859131,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.010779444128274918,
+ "learning_rate": 0.000834892657976945,
+ "loss": 3.8477983474731445,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.009554282762110233,
+ "learning_rate": 0.0008337080080246915,
+ "loss": 3.793790817260742,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.009768152609467506,
+ "learning_rate": 0.0008325239069770906,
+ "loss": 3.860384225845337,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.010801832191646099,
+ "learning_rate": 0.0008313403557881513,
+ "loss": 3.816796064376831,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.01046004518866539,
+ "learning_rate": 0.0008301573554114378,
+ "loss": 3.8064301013946533,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.009319447912275791,
+ "learning_rate": 0.0008289749068000744,
+ "loss": 3.832408905029297,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01011407095938921,
+ "learning_rate": 0.000827793010906737,
+ "loss": 3.8276009559631348,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.010393989272415638,
+ "learning_rate": 0.0008266116686836582,
+ "loss": 3.83754825592041,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.00883004255592823,
+ "learning_rate": 0.0008254308810826244,
+ "loss": 3.8315513134002686,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.008549787104129791,
+ "learning_rate": 0.0008242506490549745,
+ "loss": 3.792919158935547,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.009668390266597271,
+ "learning_rate": 0.0008230709735516006,
+ "loss": 3.8110954761505127,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.009008699096739292,
+ "learning_rate": 0.0008218918555229445,
+ "loss": 3.8598508834838867,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.008660725317895412,
+ "learning_rate": 0.0008207132959190026,
+ "loss": 3.839632511138916,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.00827972125262022,
+ "learning_rate": 0.0008195352956893172,
+ "loss": 3.8428683280944824,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.008117233403027058,
+ "learning_rate": 0.0008183578557829825,
+ "loss": 3.8434956073760986,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.008314455859363079,
+ "learning_rate": 0.0008171809771486405,
+ "loss": 3.8079686164855957,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.00865850131958723,
+ "learning_rate": 0.0008160046607344811,
+ "loss": 3.846999406814575,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.0096864253282547,
+ "learning_rate": 0.0008148289074882416,
+ "loss": 3.8227243423461914,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.00840236246585846,
+ "learning_rate": 0.0008136537183572045,
+ "loss": 3.8400344848632812,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.007451791316270828,
+ "learning_rate": 0.0008124790942881984,
+ "loss": 3.8318095207214355,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.00858070608228445,
+ "learning_rate": 0.000811305036227597,
+ "loss": 3.8232922554016113,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.00765061192214489,
+ "learning_rate": 0.0008101315451213178,
+ "loss": 3.81624174118042,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.009304090403020382,
+ "learning_rate": 0.0008089586219148213,
+ "loss": 3.8715004920959473,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.009136030450463295,
+ "learning_rate": 0.0008077862675531103,
+ "loss": 3.825037956237793,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.007460939697921276,
+ "learning_rate": 0.0008066144829807303,
+ "loss": 3.811749219894409,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.007510480936616659,
+ "learning_rate": 0.0008054432691417659,
+ "loss": 3.8313679695129395,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.008825462311506271,
+ "learning_rate": 0.0008042726269798433,
+ "loss": 3.83119535446167,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.009002318605780602,
+ "learning_rate": 0.0008031025574381281,
+ "loss": 3.8542468547821045,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.009266122244298458,
+ "learning_rate": 0.0008019330614593237,
+ "loss": 3.8328065872192383,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.008935115300118923,
+ "learning_rate": 0.0008007641399856727,
+ "loss": 3.8198719024658203,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.008021850138902664,
+ "learning_rate": 0.0007995957939589526,
+ "loss": 3.828382968902588,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.008035171777009964,
+ "learning_rate": 0.0007984280243204805,
+ "loss": 3.879868507385254,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.008553579449653625,
+ "learning_rate": 0.0007972608320111059,
+ "loss": 3.842162609100342,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.008016746491193771,
+ "learning_rate": 0.000796094217971215,
+ "loss": 3.85674786567688,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.007781418971717358,
+ "learning_rate": 0.0007949281831407276,
+ "loss": 3.7968859672546387,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.008552255108952522,
+ "learning_rate": 0.0007937627284590971,
+ "loss": 3.8305506706237793,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.007390565704554319,
+ "learning_rate": 0.0007925978548653096,
+ "loss": 3.8228843212127686,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.007611850742250681,
+ "learning_rate": 0.0007914335632978812,
+ "loss": 3.868957996368408,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.008188549429178238,
+ "learning_rate": 0.0007902698546948628,
+ "loss": 3.836078405380249,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.008198797702789307,
+ "learning_rate": 0.0007891067299938318,
+ "loss": 3.8637044429779053,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.008353965356945992,
+ "learning_rate": 0.0007879441901318977,
+ "loss": 3.833843946456909,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.0086967209354043,
+ "learning_rate": 0.0007867822360456963,
+ "loss": 3.850536823272705,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.009151408448815346,
+ "learning_rate": 0.0007856208686713949,
+ "loss": 3.8453407287597656,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.008010942488908768,
+ "learning_rate": 0.0007844600889446858,
+ "loss": 3.8679513931274414,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.007813206873834133,
+ "learning_rate": 0.0007832998978007873,
+ "loss": 3.860609531402588,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.007759213913232088,
+ "learning_rate": 0.0007821402961744448,
+ "loss": 3.803863048553467,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.008045864291489124,
+ "learning_rate": 0.0007809812849999293,
+ "loss": 3.858161449432373,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.008212368935346603,
+ "learning_rate": 0.0007798228652110347,
+ "loss": 3.8311336040496826,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.007982932962477207,
+ "learning_rate": 0.0007786650377410774,
+ "loss": 3.870339870452881,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.0071402136236429214,
+ "learning_rate": 0.0007775078035229002,
+ "loss": 3.7922263145446777,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.008564750663936138,
+ "learning_rate": 0.0007763511634888656,
+ "loss": 3.83981990814209,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.008078585378825665,
+ "learning_rate": 0.0007751951185708562,
+ "loss": 3.852938652038574,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.00801754929125309,
+ "learning_rate": 0.0007740396697002774,
+ "loss": 3.8434009552001953,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.008973612450063229,
+ "learning_rate": 0.0007728848178080535,
+ "loss": 3.8238563537597656,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.010199865326285362,
+ "learning_rate": 0.0007717305638246281,
+ "loss": 3.8891735076904297,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.009620471857488155,
+ "learning_rate": 0.0007705769086799617,
+ "loss": 3.8611769676208496,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.007981563918292522,
+ "learning_rate": 0.0007694238533035337,
+ "loss": 3.8496928215026855,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.008856567554175854,
+ "learning_rate": 0.0007682713986243415,
+ "loss": 3.8366098403930664,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.008323253132402897,
+ "learning_rate": 0.0007671195455708952,
+ "loss": 3.8224244117736816,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.00814893003553152,
+ "learning_rate": 0.0007659682950712228,
+ "loss": 3.8611154556274414,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.009146672673523426,
+ "learning_rate": 0.0007648176480528659,
+ "loss": 3.8258843421936035,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.007573423441499472,
+ "learning_rate": 0.0007636676054428805,
+ "loss": 3.888007640838623,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.00813984964042902,
+ "learning_rate": 0.0007625181681678343,
+ "loss": 3.860306739807129,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.007936778478324413,
+ "learning_rate": 0.0007613693371538078,
+ "loss": 3.8808555603027344,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.008434669114649296,
+ "learning_rate": 0.0007602211133263956,
+ "loss": 3.854154348373413,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.007857564836740494,
+ "learning_rate": 0.0007590734976106985,
+ "loss": 3.8240737915039062,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.007989327423274517,
+ "learning_rate": 0.0007579264909313315,
+ "loss": 3.850971221923828,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.008888393640518188,
+ "learning_rate": 0.0007567800942124152,
+ "loss": 3.862267255783081,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.38022518157959,
+ "eval_runtime": 51.4686,
+ "eval_samples_per_second": 47.446,
+ "eval_steps_per_second": 1.496,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.00952722318470478,
+ "learning_rate": 0.0007556343083775832,
+ "loss": 3.7705495357513428,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01099323108792305,
+ "learning_rate": 0.0007544891343499723,
+ "loss": 3.80812406539917,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.010298571549355984,
+ "learning_rate": 0.0007533445730522295,
+ "loss": 3.782794237136841,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.008139589801430702,
+ "learning_rate": 0.000752200625406507,
+ "loss": 3.7877392768859863,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.009031252935528755,
+ "learning_rate": 0.0007510572923344628,
+ "loss": 3.8181564807891846,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.00942459236830473,
+ "learning_rate": 0.0007499145747572603,
+ "loss": 3.807425022125244,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.00922947097569704,
+ "learning_rate": 0.0007487724735955648,
+ "loss": 3.7923264503479004,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.009551790542900562,
+ "learning_rate": 0.0007476309897695485,
+ "loss": 3.790083885192871,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.008996761403977871,
+ "learning_rate": 0.0007464901241988831,
+ "loss": 3.836491107940674,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.009766182862222195,
+ "learning_rate": 0.0007453498778027433,
+ "loss": 3.7829647064208984,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.007708575110882521,
+ "learning_rate": 0.0007442102514998058,
+ "loss": 3.803400993347168,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.008606902323663235,
+ "learning_rate": 0.0007430712462082464,
+ "loss": 3.810825824737549,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.009462927468121052,
+ "learning_rate": 0.0007419328628457416,
+ "loss": 3.784748077392578,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.010283366777002811,
+ "learning_rate": 0.0007407951023294655,
+ "loss": 3.744312286376953,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.008952794596552849,
+ "learning_rate": 0.0007396579655760917,
+ "loss": 3.8098015785217285,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.008446207270026207,
+ "learning_rate": 0.0007385214535017907,
+ "loss": 3.8306760787963867,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.007988285273313522,
+ "learning_rate": 0.0007373855670222302,
+ "loss": 3.7976696491241455,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.007586437277495861,
+ "learning_rate": 0.0007362503070525718,
+ "loss": 3.785891056060791,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.00775605533272028,
+ "learning_rate": 0.0007351156745074757,
+ "loss": 3.8176379203796387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.007727691903710365,
+ "learning_rate": 0.0007339816703010948,
+ "loss": 3.791743278503418,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.008311517536640167,
+ "learning_rate": 0.0007328482953470747,
+ "loss": 3.818769693374634,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.007610874716192484,
+ "learning_rate": 0.0007317155505585556,
+ "loss": 3.76772141456604,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.0078107016161084175,
+ "learning_rate": 0.0007305834368481699,
+ "loss": 3.822758197784424,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.006906864698976278,
+ "learning_rate": 0.0007294519551280412,
+ "loss": 3.832834005355835,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.008230660110712051,
+ "learning_rate": 0.0007283211063097829,
+ "loss": 3.85087513923645,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.008201118558645248,
+ "learning_rate": 0.0007271908913044997,
+ "loss": 3.7777247428894043,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.0070146857760846615,
+ "learning_rate": 0.000726061311022787,
+ "loss": 3.837517738342285,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.0076036700047552586,
+ "learning_rate": 0.0007249323663747254,
+ "loss": 3.8358592987060547,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.007238952908664942,
+ "learning_rate": 0.0007238040582698858,
+ "loss": 3.828089952468872,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.00825269054621458,
+ "learning_rate": 0.0007226763876173259,
+ "loss": 3.806016445159912,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.008368568494915962,
+ "learning_rate": 0.0007215493553255898,
+ "loss": 3.8163394927978516,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.00728471577167511,
+ "learning_rate": 0.0007204229623027063,
+ "loss": 3.835597038269043,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.00784830842167139,
+ "learning_rate": 0.0007192972094561897,
+ "loss": 3.8050971031188965,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.007714141625910997,
+ "learning_rate": 0.0007181720976930404,
+ "loss": 3.8031606674194336,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.009128103032708168,
+ "learning_rate": 0.000717047627919739,
+ "loss": 3.77107572555542,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.008033272810280323,
+ "learning_rate": 0.000715923801042251,
+ "loss": 3.8615059852600098,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.007732565049082041,
+ "learning_rate": 0.0007148006179660235,
+ "loss": 3.8277132511138916,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.007382966112345457,
+ "learning_rate": 0.0007136780795959853,
+ "loss": 3.803431987762451,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.007517020218074322,
+ "learning_rate": 0.0007125561868365442,
+ "loss": 3.799734354019165,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.008041434921324253,
+ "learning_rate": 0.0007114349405915895,
+ "loss": 3.7848258018493652,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.007875649258494377,
+ "learning_rate": 0.0007103143417644891,
+ "loss": 3.8136563301086426,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.008213640190660954,
+ "learning_rate": 0.000709194391258089,
+ "loss": 3.8518495559692383,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.00801054760813713,
+ "learning_rate": 0.0007080750899747139,
+ "loss": 3.7956271171569824,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.008064007386565208,
+ "learning_rate": 0.0007069564388161631,
+ "loss": 3.854595184326172,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.008300521411001682,
+ "learning_rate": 0.0007058384386837154,
+ "loss": 3.8203625679016113,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.008640496991574764,
+ "learning_rate": 0.000704721090478122,
+ "loss": 3.793574810028076,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.007684200536459684,
+ "learning_rate": 0.0007036043950996105,
+ "loss": 3.8564438819885254,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.008431905880570412,
+ "learning_rate": 0.0007024883534478823,
+ "loss": 3.8271119594573975,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.008137186989188194,
+ "learning_rate": 0.0007013729664221122,
+ "loss": 3.8219337463378906,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.0076744635589420795,
+ "learning_rate": 0.0007002582349209474,
+ "loss": 3.8353049755096436,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.008388740941882133,
+ "learning_rate": 0.0006991441598425065,
+ "loss": 3.795104503631592,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.007856911979615688,
+ "learning_rate": 0.0006980307420843798,
+ "loss": 3.818411111831665,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.008048887364566326,
+ "learning_rate": 0.000696917982543628,
+ "loss": 3.86599063873291,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.007949979975819588,
+ "learning_rate": 0.0006958058821167812,
+ "loss": 3.8026742935180664,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.008162821643054485,
+ "learning_rate": 0.0006946944416998389,
+ "loss": 3.833751678466797,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.008190765976905823,
+ "learning_rate": 0.0006935836621882682,
+ "loss": 3.8164236545562744,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.009714310988783836,
+ "learning_rate": 0.000692473544477005,
+ "loss": 3.8367366790771484,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.008075987920165062,
+ "learning_rate": 0.0006913640894604498,
+ "loss": 3.8108773231506348,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.009075167588889599,
+ "learning_rate": 0.0006902552980324713,
+ "loss": 3.8444485664367676,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.008678440935909748,
+ "learning_rate": 0.0006891471710864022,
+ "loss": 3.8384461402893066,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.007705411408096552,
+ "learning_rate": 0.0006880397095150414,
+ "loss": 3.825443744659424,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.00782099086791277,
+ "learning_rate": 0.0006869329142106488,
+ "loss": 3.8331687450408936,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.007542501203715801,
+ "learning_rate": 0.0006858267860649509,
+ "loss": 3.7905991077423096,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.008315361104905605,
+ "learning_rate": 0.0006847213259691356,
+ "loss": 3.861527442932129,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.008529976941645145,
+ "learning_rate": 0.0006836165348138508,
+ "loss": 3.8332972526550293,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.008117211051285267,
+ "learning_rate": 0.0006825124134892072,
+ "loss": 3.7961089611053467,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.008409497328102589,
+ "learning_rate": 0.0006814089628847755,
+ "loss": 3.812218189239502,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.007592010777443647,
+ "learning_rate": 0.0006803061838895864,
+ "loss": 3.801304340362549,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.007820053026080132,
+ "learning_rate": 0.0006792040773921283,
+ "loss": 3.8517117500305176,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.0073044621385633945,
+ "learning_rate": 0.0006781026442803479,
+ "loss": 3.8253560066223145,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.007546546868979931,
+ "learning_rate": 0.0006770018854416522,
+ "loss": 3.800802230834961,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.008878474123775959,
+ "learning_rate": 0.0006759018017629007,
+ "loss": 3.8133859634399414,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.375829219818115,
+ "eval_runtime": 51.3756,
+ "eval_samples_per_second": 47.532,
+ "eval_steps_per_second": 1.499,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.00814585480839014,
+ "learning_rate": 0.0006748023941304115,
+ "loss": 3.7970919609069824,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.00824262760579586,
+ "learning_rate": 0.0006737036634299577,
+ "loss": 3.750131607055664,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.008675593882799149,
+ "learning_rate": 0.000672605610546767,
+ "loss": 3.791456699371338,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.009052547626197338,
+ "learning_rate": 0.00067150823636552,
+ "loss": 3.7762060165405273,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.00818016566336155,
+ "learning_rate": 0.0006704115417703514,
+ "loss": 3.750631332397461,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.00800115056335926,
+ "learning_rate": 0.0006693155276448489,
+ "loss": 3.811519145965576,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.009032433852553368,
+ "learning_rate": 0.0006682201948720509,
+ "loss": 3.7805609703063965,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.0077886078506708145,
+ "learning_rate": 0.0006671255443344474,
+ "loss": 3.749366521835327,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.008603204973042011,
+ "learning_rate": 0.0006660315769139774,
+ "loss": 3.7686312198638916,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.008361323736608028,
+ "learning_rate": 0.0006649382934920329,
+ "loss": 3.740269660949707,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.00948982872068882,
+ "learning_rate": 0.0006638456949494507,
+ "loss": 3.7874045372009277,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.008958354592323303,
+ "learning_rate": 0.0006627537821665185,
+ "loss": 3.78987455368042,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.008702470920979977,
+ "learning_rate": 0.0006616625560229704,
+ "loss": 3.751061201095581,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.008424391970038414,
+ "learning_rate": 0.000660572017397988,
+ "loss": 3.80816650390625,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.009022071026265621,
+ "learning_rate": 0.0006594821671701989,
+ "loss": 3.726067304611206,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.007974586449563503,
+ "learning_rate": 0.0006583930062176739,
+ "loss": 3.78564453125,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.007887622341513634,
+ "learning_rate": 0.0006573045354179329,
+ "loss": 3.7993059158325195,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.007782666943967342,
+ "learning_rate": 0.0006562167556479352,
+ "loss": 3.7440507411956787,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.007923880591988564,
+ "learning_rate": 0.000655129667784086,
+ "loss": 3.7920784950256348,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.007850701920688152,
+ "learning_rate": 0.0006540432727022325,
+ "loss": 3.7833633422851562,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.007855751551687717,
+ "learning_rate": 0.0006529575712776633,
+ "loss": 3.812349319458008,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.007832377217710018,
+ "learning_rate": 0.0006518725643851091,
+ "loss": 3.7815146446228027,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.006997724529355764,
+ "learning_rate": 0.0006507882528987395,
+ "loss": 3.795346975326538,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.007433670572936535,
+ "learning_rate": 0.0006497046376921648,
+ "loss": 3.841017246246338,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.0075767990201711655,
+ "learning_rate": 0.0006486217196384344,
+ "loss": 3.7913427352905273,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.007551983464509249,
+ "learning_rate": 0.0006475394996100359,
+ "loss": 3.765105962753296,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.007235546596348286,
+ "learning_rate": 0.0006464579784788943,
+ "loss": 3.7668676376342773,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.007872921414673328,
+ "learning_rate": 0.0006453771571163716,
+ "loss": 3.770536422729492,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.007598782889544964,
+ "learning_rate": 0.0006442970363932667,
+ "loss": 3.796818733215332,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.007902889512479305,
+ "learning_rate": 0.0006432176171798124,
+ "loss": 3.824531078338623,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.008964902721345425,
+ "learning_rate": 0.0006421389003456776,
+ "loss": 3.781778335571289,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.009936625137925148,
+ "learning_rate": 0.000641060886759965,
+ "loss": 3.750248670578003,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.008381214924156666,
+ "learning_rate": 0.0006399835772912106,
+ "loss": 3.797954797744751,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.007752922363579273,
+ "learning_rate": 0.000638906972807384,
+ "loss": 3.789350748062134,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.009501751512289047,
+ "learning_rate": 0.0006378310741758841,
+ "loss": 3.791079521179199,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.00936005637049675,
+ "learning_rate": 0.0006367558822635453,
+ "loss": 3.78501033782959,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.008370982483029366,
+ "learning_rate": 0.0006356813979366284,
+ "loss": 3.792647361755371,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.00782831758260727,
+ "learning_rate": 0.000634607622060827,
+ "loss": 3.772569179534912,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.00830658059567213,
+ "learning_rate": 0.0006335345555012627,
+ "loss": 3.7787656784057617,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.009015833027660847,
+ "learning_rate": 0.000632462199122486,
+ "loss": 3.787418842315674,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.008993860334157944,
+ "learning_rate": 0.0006313905537884752,
+ "loss": 3.818718433380127,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.007954884320497513,
+ "learning_rate": 0.0006303196203626344,
+ "loss": 3.8100008964538574,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.007742907386273146,
+ "learning_rate": 0.0006292493997077973,
+ "loss": 3.7786598205566406,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.008748230524361134,
+ "learning_rate": 0.0006281798926862195,
+ "loss": 3.798635482788086,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.008306438103318214,
+ "learning_rate": 0.000627111100159585,
+ "loss": 3.827213764190674,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.008027350530028343,
+ "learning_rate": 0.0006260430229889987,
+ "loss": 3.7682337760925293,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.007779776118695736,
+ "learning_rate": 0.0006249756620349931,
+ "loss": 3.7515475749969482,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.008748390711843967,
+ "learning_rate": 0.000623909018157521,
+ "loss": 3.7963085174560547,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.008587505668401718,
+ "learning_rate": 0.0006228430922159577,
+ "loss": 3.7767529487609863,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.007172323763370514,
+ "learning_rate": 0.0006217778850691007,
+ "loss": 3.79768705368042,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.007224083412438631,
+ "learning_rate": 0.000620713397575168,
+ "loss": 3.785390853881836,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.008209837600588799,
+ "learning_rate": 0.0006196496305917987,
+ "loss": 3.809828281402588,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.007760563865303993,
+ "learning_rate": 0.0006185865849760491,
+ "loss": 3.838716506958008,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.007362260017544031,
+ "learning_rate": 0.0006175242615843977,
+ "loss": 3.7840137481689453,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.008719335310161114,
+ "learning_rate": 0.0006164626612727381,
+ "loss": 3.8228893280029297,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.00856029987335205,
+ "learning_rate": 0.0006154017848963826,
+ "loss": 3.80448317527771,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.008078506216406822,
+ "learning_rate": 0.0006143416333100606,
+ "loss": 3.7881453037261963,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.007957885973155499,
+ "learning_rate": 0.0006132822073679164,
+ "loss": 3.814870834350586,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.009362339973449707,
+ "learning_rate": 0.0006122235079235114,
+ "loss": 3.798335075378418,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.008272632025182247,
+ "learning_rate": 0.0006111655358298194,
+ "loss": 3.8212976455688477,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.007548908703029156,
+ "learning_rate": 0.0006101082919392301,
+ "loss": 3.8312830924987793,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.007543765474110842,
+ "learning_rate": 0.0006090517771035457,
+ "loss": 3.8302183151245117,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.007520351093262434,
+ "learning_rate": 0.000607995992173981,
+ "loss": 3.7898802757263184,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.007168850861489773,
+ "learning_rate": 0.000606940938001163,
+ "loss": 3.8078577518463135,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.007420001085847616,
+ "learning_rate": 0.0006058866154351296,
+ "loss": 3.799571990966797,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.007621012162417173,
+ "learning_rate": 0.0006048330253253303,
+ "loss": 3.833843946456909,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.007661828305572271,
+ "learning_rate": 0.0006037801685206222,
+ "loss": 3.820556640625,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.007657836657017469,
+ "learning_rate": 0.0006027280458692738,
+ "loss": 3.798994541168213,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.007722760550677776,
+ "learning_rate": 0.0006016766582189613,
+ "loss": 3.8589251041412354,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.00789247453212738,
+ "learning_rate": 0.0006006260064167685,
+ "loss": 3.8673787117004395,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.009137817658483982,
+ "learning_rate": 0.0005995760913091874,
+ "loss": 3.801938056945801,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.008374894969165325,
+ "learning_rate": 0.0005985269137421141,
+ "loss": 3.820045232772827,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.3730878829956055,
+ "eval_runtime": 48.1607,
+ "eval_samples_per_second": 50.705,
+ "eval_steps_per_second": 1.599,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.009492472745478153,
+ "learning_rate": 0.0005974784745608536,
+ "loss": 3.738617420196533,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.010017318651080132,
+ "learning_rate": 0.0005964307746101135,
+ "loss": 3.7551231384277344,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.008958659134805202,
+ "learning_rate": 0.000595383814734007,
+ "loss": 3.7739763259887695,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.008034365251660347,
+ "learning_rate": 0.0005943375957760505,
+ "loss": 3.7678627967834473,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.008489780128002167,
+ "learning_rate": 0.000593292118579164,
+ "loss": 3.720778465270996,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.009107839316129684,
+ "learning_rate": 0.0005922473839856699,
+ "loss": 3.7224440574645996,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.009295578114688396,
+ "learning_rate": 0.0005912033928372906,
+ "loss": 3.7695257663726807,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.007896862924098969,
+ "learning_rate": 0.0005901601459751532,
+ "loss": 3.742567300796509,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.008227395825088024,
+ "learning_rate": 0.0005891176442397808,
+ "loss": 3.7547202110290527,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.009326796047389507,
+ "learning_rate": 0.0005880758884710992,
+ "loss": 3.724808931350708,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.007786506786942482,
+ "learning_rate": 0.0005870348795084319,
+ "loss": 3.7403640747070312,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.00821041688323021,
+ "learning_rate": 0.0005859946181905009,
+ "loss": 3.770012617111206,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.009544595144689083,
+ "learning_rate": 0.000584955105355427,
+ "loss": 3.7378525733947754,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.008973932825028896,
+ "learning_rate": 0.0005839163418407254,
+ "loss": 3.7184295654296875,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.009023688733577728,
+ "learning_rate": 0.00058287832848331,
+ "loss": 3.7388229370117188,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.008151066489517689,
+ "learning_rate": 0.0005818410661194889,
+ "loss": 3.7632875442504883,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.008495030924677849,
+ "learning_rate": 0.0005808045555849668,
+ "loss": 3.8114736080169678,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.008739210665225983,
+ "learning_rate": 0.0005797687977148394,
+ "loss": 3.7647829055786133,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.008062886074185371,
+ "learning_rate": 0.0005787337933435999,
+ "loss": 3.782172441482544,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.007615398149937391,
+ "learning_rate": 0.0005776995433051326,
+ "loss": 3.7564642429351807,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.007938290014863014,
+ "learning_rate": 0.0005766660484327133,
+ "loss": 3.756208896636963,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.008517666719853878,
+ "learning_rate": 0.0005756333095590102,
+ "loss": 3.76729416847229,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.007426336407661438,
+ "learning_rate": 0.0005746013275160827,
+ "loss": 3.74001407623291,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.007664753124117851,
+ "learning_rate": 0.0005735701031353804,
+ "loss": 3.7819876670837402,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.007903067395091057,
+ "learning_rate": 0.0005725396372477414,
+ "loss": 3.781219959259033,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.007251975126564503,
+ "learning_rate": 0.0005715099306833931,
+ "loss": 3.763620615005493,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.007422176189720631,
+ "learning_rate": 0.0005704809842719533,
+ "loss": 3.775099754333496,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.009161600843071938,
+ "learning_rate": 0.0005694527988424239,
+ "loss": 3.792268991470337,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.008330529555678368,
+ "learning_rate": 0.0005684253752231958,
+ "loss": 3.7967824935913086,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.0073860278353095055,
+ "learning_rate": 0.0005673987142420458,
+ "loss": 3.76505970954895,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.008462096564471722,
+ "learning_rate": 0.0005663728167261363,
+ "loss": 3.740450143814087,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.00838128849864006,
+ "learning_rate": 0.0005653476835020138,
+ "loss": 3.7863235473632812,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.007381014991551638,
+ "learning_rate": 0.0005643233153956092,
+ "loss": 3.772179126739502,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.007363761775195599,
+ "learning_rate": 0.0005632997132322392,
+ "loss": 3.7793867588043213,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.008684969507157803,
+ "learning_rate": 0.0005622768778366001,
+ "loss": 3.7736339569091797,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.007934059016406536,
+ "learning_rate": 0.0005612548100327722,
+ "loss": 3.8040053844451904,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.007136524189263582,
+ "learning_rate": 0.0005602335106442172,
+ "loss": 3.7589468955993652,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.00789916142821312,
+ "learning_rate": 0.0005592129804937779,
+ "loss": 3.7953195571899414,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.007919179275631905,
+ "learning_rate": 0.0005581932204036761,
+ "loss": 3.7539429664611816,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.007734223268926144,
+ "learning_rate": 0.0005571742311955147,
+ "loss": 3.7620325088500977,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.007005508057773113,
+ "learning_rate": 0.0005561560136902749,
+ "loss": 3.762538433074951,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.008137146942317486,
+ "learning_rate": 0.0005551385687083161,
+ "loss": 3.773789405822754,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.007310070563107729,
+ "learning_rate": 0.0005541218970693757,
+ "loss": 3.8077168464660645,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.007335536181926727,
+ "learning_rate": 0.0005531059995925661,
+ "loss": 3.7761921882629395,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.008022893220186234,
+ "learning_rate": 0.00055209087709638,
+ "loss": 3.7769880294799805,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.008027128875255585,
+ "learning_rate": 0.0005510765303986814,
+ "loss": 3.7709951400756836,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.00697955721989274,
+ "learning_rate": 0.0005500629603167118,
+ "loss": 3.7855210304260254,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.007455077022314072,
+ "learning_rate": 0.0005490501676670861,
+ "loss": 3.7850632667541504,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.0069657545536756516,
+ "learning_rate": 0.0005480381532657932,
+ "loss": 3.742671251296997,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.007609451189637184,
+ "learning_rate": 0.0005470269179281953,
+ "loss": 3.793771266937256,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.007437967229634523,
+ "learning_rate": 0.0005460164624690255,
+ "loss": 3.778846263885498,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.007123028859496117,
+ "learning_rate": 0.00054500678770239,
+ "loss": 3.768375873565674,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.008176218718290329,
+ "learning_rate": 0.0005439978944417654,
+ "loss": 3.82069730758667,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.006943001411855221,
+ "learning_rate": 0.000542989783499999,
+ "loss": 3.7731552124023438,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.0072205523028969765,
+ "learning_rate": 0.0005419824556893069,
+ "loss": 3.7724058628082275,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.007656400091946125,
+ "learning_rate": 0.0005409759118212758,
+ "loss": 3.780989646911621,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.006972154602408409,
+ "learning_rate": 0.0005399701527068599,
+ "loss": 3.7971787452697754,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.007674331311136484,
+ "learning_rate": 0.0005389651791563806,
+ "loss": 3.798673391342163,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.006952045019716024,
+ "learning_rate": 0.0005379609919795268,
+ "loss": 3.7982821464538574,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.007882075384259224,
+ "learning_rate": 0.0005369575919853545,
+ "loss": 3.825962781906128,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.007143386639654636,
+ "learning_rate": 0.0005359549799822854,
+ "loss": 3.826598644256592,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.007587207015603781,
+ "learning_rate": 0.0005349531567781041,
+ "loss": 3.80623197555542,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.007677781395614147,
+ "learning_rate": 0.0005339521231799632,
+ "loss": 3.7987256050109863,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.007372343447059393,
+ "learning_rate": 0.0005329518799943773,
+ "loss": 3.7733969688415527,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.00861304346472025,
+ "learning_rate": 0.0005319524280272231,
+ "loss": 3.7903685569763184,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.00816237274557352,
+ "learning_rate": 0.0005309537680837418,
+ "loss": 3.77058744430542,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.007624274119734764,
+ "learning_rate": 0.0005299559009685356,
+ "loss": 3.787982940673828,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.007416521664708853,
+ "learning_rate": 0.0005289588274855683,
+ "loss": 3.7991762161254883,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.007461614441126585,
+ "learning_rate": 0.0005279625484381632,
+ "loss": 3.8036863803863525,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.007766335736960173,
+ "learning_rate": 0.0005269670646290043,
+ "loss": 3.809243679046631,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.0073394919745624065,
+ "learning_rate": 0.0005259723768601365,
+ "loss": 3.756606101989746,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.008771199733018875,
+ "learning_rate": 0.0005249784859329602,
+ "loss": 3.773951768875122,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.370637893676758,
+ "eval_runtime": 49.4904,
+ "eval_samples_per_second": 49.343,
+ "eval_steps_per_second": 1.556,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.008164486847817898,
+ "learning_rate": 0.0005239853926482358,
+ "loss": 3.7619595527648926,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.007980132475495338,
+ "learning_rate": 0.0005229930978060806,
+ "loss": 3.7398204803466797,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.0077056242153048515,
+ "learning_rate": 0.0005220016022059692,
+ "loss": 3.699580669403076,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.0074698119424283504,
+ "learning_rate": 0.0005210109066467308,
+ "loss": 3.7299928665161133,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.008384468965232372,
+ "learning_rate": 0.0005200210119265512,
+ "loss": 3.704989433288574,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.008066141977906227,
+ "learning_rate": 0.0005190319188429709,
+ "loss": 3.751697540283203,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.007213297300040722,
+ "learning_rate": 0.0005180436281928841,
+ "loss": 3.740692377090454,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.007343464996665716,
+ "learning_rate": 0.0005170561407725396,
+ "loss": 3.7060189247131348,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.00783725269138813,
+ "learning_rate": 0.0005160694573775363,
+ "loss": 3.7414426803588867,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.007640232797712088,
+ "learning_rate": 0.0005150835788028294,
+ "loss": 3.7451486587524414,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.008129071444272995,
+ "learning_rate": 0.0005140985058427218,
+ "loss": 3.734839916229248,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.006920623127371073,
+ "learning_rate": 0.0005131142392908697,
+ "loss": 3.7138614654541016,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.00798740889877081,
+ "learning_rate": 0.0005121307799402787,
+ "loss": 3.76496958732605,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.006689135450869799,
+ "learning_rate": 0.0005111481285833045,
+ "loss": 3.7707948684692383,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.007544311694800854,
+ "learning_rate": 0.0005101662860116516,
+ "loss": 3.7393479347229004,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.007487910334020853,
+ "learning_rate": 0.0005091852530163717,
+ "loss": 3.7486305236816406,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.00772349676117301,
+ "learning_rate": 0.0005082050303878673,
+ "loss": 3.753629207611084,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.007883663289248943,
+ "learning_rate": 0.0005072256189158844,
+ "loss": 3.7497611045837402,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.008153906092047691,
+ "learning_rate": 0.0005062470193895178,
+ "loss": 3.7459521293640137,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.007083641365170479,
+ "learning_rate": 0.0005052692325972075,
+ "loss": 3.725599765777588,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.007102683652192354,
+ "learning_rate": 0.0005042922593267385,
+ "loss": 3.7724428176879883,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.008104706183075905,
+ "learning_rate": 0.0005033161003652409,
+ "loss": 3.7386724948883057,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.0073765418492257595,
+ "learning_rate": 0.0005023407564991878,
+ "loss": 3.7478132247924805,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.007110628765076399,
+ "learning_rate": 0.0005013662285143963,
+ "loss": 3.771925687789917,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.007323848083615303,
+ "learning_rate": 0.000500392517196026,
+ "loss": 3.7460904121398926,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.007830807007849216,
+ "learning_rate": 0.0004994196233285786,
+ "loss": 3.751222610473633,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.007710007485002279,
+ "learning_rate": 0.0004984475476958973,
+ "loss": 3.739858627319336,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.007293973118066788,
+ "learning_rate": 0.0004974762910811655,
+ "loss": 3.7410120964050293,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.0073824538849294186,
+ "learning_rate": 0.0004965058542669078,
+ "loss": 3.711397647857666,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.007156423293054104,
+ "learning_rate": 0.0004955362380349866,
+ "loss": 3.7881593704223633,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.007705818396061659,
+ "learning_rate": 0.0004945674431666047,
+ "loss": 3.7639269828796387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.0075300210155546665,
+ "learning_rate": 0.0004935994704423026,
+ "loss": 3.7562999725341797,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.007137116510421038,
+ "learning_rate": 0.000492632320641959,
+ "loss": 3.7659149169921875,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.007225630804896355,
+ "learning_rate": 0.0004916659945447878,
+ "loss": 3.7625155448913574,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.008027992211282253,
+ "learning_rate": 0.0004907004929293408,
+ "loss": 3.746478796005249,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.007582477759569883,
+ "learning_rate": 0.0004897358165735066,
+ "loss": 3.759408950805664,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.007575645577162504,
+ "learning_rate": 0.000488771966254506,
+ "loss": 3.7430644035339355,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.008404700085520744,
+ "learning_rate": 0.0004878089427488963,
+ "loss": 3.7727198600769043,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.007118664216250181,
+ "learning_rate": 0.00048684674683256815,
+ "loss": 3.747460126876831,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.007220778148621321,
+ "learning_rate": 0.00048588537928074584,
+ "loss": 3.746413230895996,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.0076194945722818375,
+ "learning_rate": 0.0004849248408679849,
+ "loss": 3.7808125019073486,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.007285538129508495,
+ "learning_rate": 0.00048396513236817387,
+ "loss": 3.7750301361083984,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.007698678877204657,
+ "learning_rate": 0.00048300625455453383,
+ "loss": 3.7474594116210938,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.007687424309551716,
+ "learning_rate": 0.0004820482081996141,
+ "loss": 3.7929720878601074,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.007363605313003063,
+ "learning_rate": 0.0004810909940752959,
+ "loss": 3.75911021232605,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.007791243493556976,
+ "learning_rate": 0.00048013461295278836,
+ "loss": 3.7567291259765625,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.007375436834990978,
+ "learning_rate": 0.0004791790656026323,
+ "loss": 3.7784414291381836,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.007330593653023243,
+ "learning_rate": 0.0004782243527946934,
+ "loss": 3.7533810138702393,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.006926248781383038,
+ "learning_rate": 0.00047727047529816694,
+ "loss": 3.7795910835266113,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.007102136500179768,
+ "learning_rate": 0.00047631743388157496,
+ "loss": 3.812549114227295,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.006676106248050928,
+ "learning_rate": 0.00047536522931276555,
+ "loss": 3.7578792572021484,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.006926767062395811,
+ "learning_rate": 0.0004744138623589134,
+ "loss": 3.782416820526123,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.006894777063280344,
+ "learning_rate": 0.000473463333786516,
+ "loss": 3.7840042114257812,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.007393381092697382,
+ "learning_rate": 0.00047251364436139923,
+ "loss": 3.7532780170440674,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.006510896608233452,
+ "learning_rate": 0.0004715647948487092,
+ "loss": 3.7618460655212402,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.007283828221261501,
+ "learning_rate": 0.0004706167860129171,
+ "loss": 3.768498420715332,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.00655644154176116,
+ "learning_rate": 0.0004696696186178168,
+ "loss": 3.7724053859710693,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.007338619790971279,
+ "learning_rate": 0.0004687232934265239,
+ "loss": 3.7814595699310303,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.006839970126748085,
+ "learning_rate": 0.0004677778112014758,
+ "loss": 3.771808624267578,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.006776005961000919,
+ "learning_rate": 0.0004668331727044297,
+ "loss": 3.786924123764038,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.007011357229202986,
+ "learning_rate": 0.00046588937869646415,
+ "loss": 3.787672996520996,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.006482086144387722,
+ "learning_rate": 0.0004649464299379771,
+ "loss": 3.7794556617736816,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.0069654518738389015,
+ "learning_rate": 0.000464004327188685,
+ "loss": 3.7520360946655273,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.006526473443955183,
+ "learning_rate": 0.0004630630712076235,
+ "loss": 3.734144926071167,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.006822322495281696,
+ "learning_rate": 0.00046212266275314504,
+ "loss": 3.7614707946777344,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.006827171426266432,
+ "learning_rate": 0.00046118310258292045,
+ "loss": 3.755218029022217,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.00706718722358346,
+ "learning_rate": 0.00046024439145393546,
+ "loss": 3.7397823333740234,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.00689381267875433,
+ "learning_rate": 0.00045930653012249314,
+ "loss": 3.758922576904297,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.007069882471114397,
+ "learning_rate": 0.0004583695193442113,
+ "loss": 3.759767532348633,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.006889182608574629,
+ "learning_rate": 0.0004574333598740228,
+ "loss": 3.7555460929870605,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.006866258103400469,
+ "learning_rate": 0.00045649805246617475,
+ "loss": 3.762068271636963,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.008696220815181732,
+ "learning_rate": 0.000455563597874226,
+ "loss": 3.7110466957092285,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.369833469390869,
+ "eval_runtime": 48.7852,
+ "eval_samples_per_second": 50.056,
+ "eval_steps_per_second": 1.578,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.0077184769324958324,
+ "learning_rate": 0.0004546299968510516,
+ "loss": 3.704669713973999,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.00846019759774208,
+ "learning_rate": 0.0004536972501488358,
+ "loss": 3.701132297515869,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.008113945834338665,
+ "learning_rate": 0.00045276535851907613,
+ "loss": 3.699582099914551,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.0072370911948382854,
+ "learning_rate": 0.0004518343227125807,
+ "loss": 3.698126792907715,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.00817187875509262,
+ "learning_rate": 0.0004509041434794685,
+ "loss": 3.738393783569336,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.007495569530874491,
+ "learning_rate": 0.00044997482156916845,
+ "loss": 3.727083683013916,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.007855897769331932,
+ "learning_rate": 0.0004490463577304171,
+ "loss": 3.716137170791626,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.007817867211997509,
+ "learning_rate": 0.0004481187527112628,
+ "loss": 3.750791072845459,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.007712380029261112,
+ "learning_rate": 0.00044719200725905887,
+ "loss": 3.720554828643799,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.00859921332448721,
+ "learning_rate": 0.00044626612212046816,
+ "loss": 3.742496967315674,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.007692081853747368,
+ "learning_rate": 0.0004453410980414582,
+ "loss": 3.703803300857544,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.007734883576631546,
+ "learning_rate": 0.00044441693576730565,
+ "loss": 3.7371301651000977,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.009022004902362823,
+ "learning_rate": 0.0004434936360425916,
+ "loss": 3.727294683456421,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.008058780804276466,
+ "learning_rate": 0.00044257119961120067,
+ "loss": 3.7583885192871094,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.00698013324290514,
+ "learning_rate": 0.00044164962721632376,
+ "loss": 3.7312440872192383,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.007999536581337452,
+ "learning_rate": 0.00044072891960045523,
+ "loss": 3.6972708702087402,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.008125107735395432,
+ "learning_rate": 0.0004398090775053929,
+ "loss": 3.7123398780822754,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.006736223120242357,
+ "learning_rate": 0.0004388901016722357,
+ "loss": 3.749678134918213,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.009154478088021278,
+ "learning_rate": 0.000437971992841387,
+ "loss": 3.722728967666626,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.008070871233940125,
+ "learning_rate": 0.00043705475175255087,
+ "loss": 3.7297017574310303,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.007498566992580891,
+ "learning_rate": 0.0004361383791447307,
+ "loss": 3.7286131381988525,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.007642887998372316,
+ "learning_rate": 0.00043522287575623183,
+ "loss": 3.750898838043213,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.007800443097949028,
+ "learning_rate": 0.0004343082423246588,
+ "loss": 3.719512462615967,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.007393732666969299,
+ "learning_rate": 0.0004333944795869159,
+ "loss": 3.7136006355285645,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.008559555746614933,
+ "learning_rate": 0.0004324815882792042,
+ "loss": 3.7087020874023438,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.007379850372672081,
+ "learning_rate": 0.0004315695691370241,
+ "loss": 3.7638165950775146,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.007374506443738937,
+ "learning_rate": 0.00043065842289517303,
+ "loss": 3.7208821773529053,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.00755207147449255,
+ "learning_rate": 0.0004297481502877451,
+ "loss": 3.738490581512451,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.007752610836178064,
+ "learning_rate": 0.0004288387520481306,
+ "loss": 3.7401833534240723,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.006907653529196978,
+ "learning_rate": 0.0004279302289090151,
+ "loss": 3.7754998207092285,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.007771312724798918,
+ "learning_rate": 0.0004270225816023797,
+ "loss": 3.73626446723938,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.007395982276648283,
+ "learning_rate": 0.0004261158108594987,
+ "loss": 3.7159175872802734,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.007196040358394384,
+ "learning_rate": 0.0004252099174109411,
+ "loss": 3.750246047973633,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.008241173811256886,
+ "learning_rate": 0.0004243049019865692,
+ "loss": 3.70859694480896,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.0072947596199810505,
+ "learning_rate": 0.00042340076531553755,
+ "loss": 3.7385268211364746,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.00782464537769556,
+ "learning_rate": 0.00042249750812629297,
+ "loss": 3.75885272026062,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.007813232019543648,
+ "learning_rate": 0.0004215951311465722,
+ "loss": 3.71810245513916,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.006900500506162643,
+ "learning_rate": 0.00042069363510340604,
+ "loss": 3.7365174293518066,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.007812316995114088,
+ "learning_rate": 0.0004197930207231119,
+ "loss": 3.774794101715088,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.007018622010946274,
+ "learning_rate": 0.00041889328873129933,
+ "loss": 3.7112040519714355,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.006895892322063446,
+ "learning_rate": 0.0004179944398528659,
+ "loss": 3.7497243881225586,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.0071940235793590546,
+ "learning_rate": 0.0004170964748119979,
+ "loss": 3.7566123008728027,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.006776201073080301,
+ "learning_rate": 0.0004161993943321702,
+ "loss": 3.7683968544006348,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.006886148825287819,
+ "learning_rate": 0.0004153031991361428,
+ "loss": 3.759763240814209,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.006466342136263847,
+ "learning_rate": 0.0004144078899459659,
+ "loss": 3.736006498336792,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.007055724039673805,
+ "learning_rate": 0.0004135134674829725,
+ "loss": 3.70645809173584,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.006948428228497505,
+ "learning_rate": 0.00041261993246778303,
+ "loss": 3.7365002632141113,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.006491438951343298,
+ "learning_rate": 0.00041172728562030255,
+ "loss": 3.735001564025879,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.006851755082607269,
+ "learning_rate": 0.0004108355276597209,
+ "loss": 3.720893383026123,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.007266475819051266,
+ "learning_rate": 0.00040994465930451114,
+ "loss": 3.7746100425720215,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.007112737279385328,
+ "learning_rate": 0.00040905468127242926,
+ "loss": 3.7441720962524414,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.006514255423098803,
+ "learning_rate": 0.000408165594280515,
+ "loss": 3.7333898544311523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.007172788958996534,
+ "learning_rate": 0.0004072773990450896,
+ "loss": 3.7675576210021973,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.00684684282168746,
+ "learning_rate": 0.0004063900962817564,
+ "loss": 3.7141308784484863,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.006578145083039999,
+ "learning_rate": 0.00040550368670539777,
+ "loss": 3.7256317138671875,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.007047492545098066,
+ "learning_rate": 0.00040461817103017926,
+ "loss": 3.7618556022644043,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.007036720868200064,
+ "learning_rate": 0.00040373354996954487,
+ "loss": 3.7397944927215576,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.00745626725256443,
+ "learning_rate": 0.00040284982423621714,
+ "loss": 3.7831153869628906,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.007454817183315754,
+ "learning_rate": 0.00040196699454219797,
+ "loss": 3.7409911155700684,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.0075650583021342754,
+ "learning_rate": 0.00040108506159876756,
+ "loss": 3.7074005603790283,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.006908242590725422,
+ "learning_rate": 0.0004002040261164838,
+ "loss": 3.767744779586792,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.0072857895866036415,
+ "learning_rate": 0.0003993238888051801,
+ "loss": 3.7349488735198975,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.0072942771948874,
+ "learning_rate": 0.00039844465037396716,
+ "loss": 3.75968861579895,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.0071496241725981236,
+ "learning_rate": 0.00039756631153123296,
+ "loss": 3.8237380981445312,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.006803249940276146,
+ "learning_rate": 0.00039668887298463783,
+ "loss": 3.7468650341033936,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.007433352991938591,
+ "learning_rate": 0.00039581233544111867,
+ "loss": 3.783236503601074,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.006633558310568333,
+ "learning_rate": 0.0003949366996068858,
+ "loss": 3.75020694732666,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.007415154483169317,
+ "learning_rate": 0.00039406196618742376,
+ "loss": 3.7545979022979736,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.006997843272984028,
+ "learning_rate": 0.0003931881358874886,
+ "loss": 3.7404396533966064,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.007417330518364906,
+ "learning_rate": 0.0003923152094111095,
+ "loss": 3.7575364112854004,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.006936701480299234,
+ "learning_rate": 0.0003914431874615889,
+ "loss": 3.771845817565918,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.008293208666145802,
+ "learning_rate": 0.00039057207074149795,
+ "loss": 3.746669292449951,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.370476245880127,
+ "eval_runtime": 50.9368,
+ "eval_samples_per_second": 47.942,
+ "eval_steps_per_second": 1.512,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.0077544888481497765,
+ "learning_rate": 0.0003897018599526799,
+ "loss": 3.6915597915649414,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.008674043230712414,
+ "learning_rate": 0.000388832555796248,
+ "loss": 3.67698073387146,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.00724575063213706,
+ "learning_rate": 0.00038796415897258523,
+ "loss": 3.6838502883911133,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.00849415734410286,
+ "learning_rate": 0.00038709667018134246,
+ "loss": 3.69064998626709,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.00769044877961278,
+ "learning_rate": 0.0003862300901214403,
+ "loss": 3.670502185821533,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.007846124470233917,
+ "learning_rate": 0.00038536441949106647,
+ "loss": 3.767771005630493,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.007555077783763409,
+ "learning_rate": 0.000384499658987676,
+ "loss": 3.7001633644104004,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.007834251970052719,
+ "learning_rate": 0.0003836358093079915,
+ "loss": 3.704331874847412,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.007608097977936268,
+ "learning_rate": 0.00038277287114799974,
+ "loss": 3.7108519077301025,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.007420446258038282,
+ "learning_rate": 0.0003819108452029561,
+ "loss": 3.697843313217163,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.008091001771390438,
+ "learning_rate": 0.000381049732167378,
+ "loss": 3.7380661964416504,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.007265515625476837,
+ "learning_rate": 0.00038018953273504907,
+ "loss": 3.680748462677002,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.007735357619822025,
+ "learning_rate": 0.00037933024759901636,
+ "loss": 3.6648077964782715,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.008188895881175995,
+ "learning_rate": 0.0003784718774515907,
+ "loss": 3.706609010696411,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.00816027820110321,
+ "learning_rate": 0.00037761442298434554,
+ "loss": 3.7368297576904297,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.006931936834007502,
+ "learning_rate": 0.00037675788488811566,
+ "loss": 3.7121026515960693,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.007615855894982815,
+ "learning_rate": 0.00037590226385299897,
+ "loss": 3.7379372119903564,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.0073505607433617115,
+ "learning_rate": 0.0003750475605683535,
+ "loss": 3.7330894470214844,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.007151979021728039,
+ "learning_rate": 0.00037419377572279897,
+ "loss": 3.713761568069458,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.007310053799301386,
+ "learning_rate": 0.00037334091000421326,
+ "loss": 3.705451011657715,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.0072968751192092896,
+ "learning_rate": 0.0003724889640997357,
+ "loss": 3.7550907135009766,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.007197601720690727,
+ "learning_rate": 0.00037163793869576404,
+ "loss": 3.7346463203430176,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.007823911495506763,
+ "learning_rate": 0.00037078783447795307,
+ "loss": 3.7257699966430664,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.007118315435945988,
+ "learning_rate": 0.0003699386521312171,
+ "loss": 3.7014496326446533,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.0069230347871780396,
+ "learning_rate": 0.00036909039233972666,
+ "loss": 3.6709237098693848,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.007836844772100449,
+ "learning_rate": 0.00036824305578691,
+ "loss": 3.695725440979004,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.007123398594558239,
+ "learning_rate": 0.00036739664315544956,
+ "loss": 3.7278594970703125,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.007792143151164055,
+ "learning_rate": 0.00036655115512728586,
+ "loss": 3.7291927337646484,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.007245976477861404,
+ "learning_rate": 0.00036570659238361345,
+ "loss": 3.7183046340942383,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.007778054103255272,
+ "learning_rate": 0.0003648629556048804,
+ "loss": 3.7485761642456055,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.007539471611380577,
+ "learning_rate": 0.00036402024547079007,
+ "loss": 3.7273335456848145,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.007554007228463888,
+ "learning_rate": 0.0003631784626602987,
+ "loss": 3.684359073638916,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.007862667553126812,
+ "learning_rate": 0.00036233760785161606,
+ "loss": 3.7223072052001953,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.00869659148156643,
+ "learning_rate": 0.0003614976817222028,
+ "loss": 3.7475168704986572,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.00696196174249053,
+ "learning_rate": 0.00036065868494877226,
+ "loss": 3.7773914337158203,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.008952745236456394,
+ "learning_rate": 0.00035982061820729005,
+ "loss": 3.6761155128479004,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.008341601118445396,
+ "learning_rate": 0.00035898348217297034,
+ "loss": 3.746553421020508,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.007268146146088839,
+ "learning_rate": 0.00035814727752027885,
+ "loss": 3.6921701431274414,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.007401647046208382,
+ "learning_rate": 0.00035731200492293064,
+ "loss": 3.722433090209961,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.006969883106648922,
+ "learning_rate": 0.00035647766505389006,
+ "loss": 3.7449448108673096,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.007616426330059767,
+ "learning_rate": 0.00035564425858536895,
+ "loss": 3.725113868713379,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.006883042864501476,
+ "learning_rate": 0.00035481178618882814,
+ "loss": 3.717106342315674,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.007106961216777563,
+ "learning_rate": 0.00035398024853497574,
+ "loss": 3.74045467376709,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.007402827963232994,
+ "learning_rate": 0.00035314964629376664,
+ "loss": 3.7288095951080322,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.006770852487534285,
+ "learning_rate": 0.00035231998013440227,
+ "loss": 3.7663369178771973,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.007622968405485153,
+ "learning_rate": 0.0003514912507253287,
+ "loss": 3.747084140777588,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.006876434665173292,
+ "learning_rate": 0.00035066345873423975,
+ "loss": 3.7591025829315186,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.007071215659379959,
+ "learning_rate": 0.00034983660482807116,
+ "loss": 3.7583885192871094,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.007130554411560297,
+ "learning_rate": 0.00034901068967300465,
+ "loss": 3.6950902938842773,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.006981188431382179,
+ "learning_rate": 0.00034818571393446524,
+ "loss": 3.7390527725219727,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.0067793820053339005,
+ "learning_rate": 0.0003473616782771209,
+ "loss": 3.703673839569092,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.00674026133492589,
+ "learning_rate": 0.0003465385833648827,
+ "loss": 3.6988182067871094,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.006553220096975565,
+ "learning_rate": 0.0003457164298609021,
+ "loss": 3.7367048263549805,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.007079749833792448,
+ "learning_rate": 0.000344895218427575,
+ "loss": 3.773226022720337,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.006091540213674307,
+ "learning_rate": 0.00034407494972653547,
+ "loss": 3.747929096221924,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.006783034652471542,
+ "learning_rate": 0.00034325562441865935,
+ "loss": 3.7154481410980225,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.006683558225631714,
+ "learning_rate": 0.0003424372431640625,
+ "loss": 3.751605987548828,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.0066695502027869225,
+ "learning_rate": 0.0003416198066220997,
+ "loss": 3.7561278343200684,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.006601935718208551,
+ "learning_rate": 0.0003408033154513655,
+ "loss": 3.754878520965576,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.006911450996994972,
+ "learning_rate": 0.00033998777030969085,
+ "loss": 3.69533109664917,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.0070345234125852585,
+ "learning_rate": 0.0003391731718541465,
+ "loss": 3.7328338623046875,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.007238153368234634,
+ "learning_rate": 0.00033835952074103964,
+ "loss": 3.7370123863220215,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.007295276038348675,
+ "learning_rate": 0.0003375468176259142,
+ "loss": 3.7123947143554688,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.006471444386988878,
+ "learning_rate": 0.0003367350631635506,
+ "loss": 3.73347806930542,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.007136075291782618,
+ "learning_rate": 0.00033592425800796447,
+ "loss": 3.7337799072265625,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.0065771411173045635,
+ "learning_rate": 0.00033511440281240725,
+ "loss": 3.733506441116333,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.006566311698406935,
+ "learning_rate": 0.0003343054982293643,
+ "loss": 3.7325632572174072,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.006389300804585218,
+ "learning_rate": 0.00033349754491055535,
+ "loss": 3.7027206420898438,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.006425539031624794,
+ "learning_rate": 0.00033269054350693386,
+ "loss": 3.751466989517212,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.006735894829034805,
+ "learning_rate": 0.0003318844946686865,
+ "loss": 3.731363296508789,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.0061791036278009415,
+ "learning_rate": 0.00033107939904523264,
+ "loss": 3.7592201232910156,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.007923831231892109,
+ "learning_rate": 0.0003302752572852219,
+ "loss": 3.7247159481048584,
+ "step": 2880
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.2278453445328896e+18,
+ "train_batch_size": 16,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/baseline/checkpoint-2880/training_args.bin b/runs/baseline/checkpoint-2880/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7
--- /dev/null
+++ b/runs/baseline/checkpoint-2880/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97
+size 5265
diff --git a/runs/baseline/checkpoint-3240/chat_template.jinja b/runs/baseline/checkpoint-3240/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/baseline/checkpoint-3240/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/baseline/checkpoint-3240/config.json b/runs/baseline/checkpoint-3240/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/baseline/checkpoint-3240/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/baseline/checkpoint-3240/generation_config.json b/runs/baseline/checkpoint-3240/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/baseline/checkpoint-3240/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/baseline/checkpoint-3240/model.safetensors b/runs/baseline/checkpoint-3240/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..9da137a2f1d4f3044d91a20bc418ad19587c3a50
--- /dev/null
+++ b/runs/baseline/checkpoint-3240/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fb208b9b92f9018d170f14ed359a3e6c33dc4579694afd0e56e77fb2daf1d8d3
+size 583356232
diff --git a/runs/baseline/checkpoint-3240/optimizer.pt b/runs/baseline/checkpoint-3240/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..f5ae832414798980d79fc475e26a5035045e7b0e
--- /dev/null
+++ b/runs/baseline/checkpoint-3240/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:678f6eb482541f0d4bd629fe6b67a25e2edb11e98bee028d66fb82583a5c94d2
+size 1166825803
diff --git a/runs/baseline/checkpoint-3240/rng_state_0.pth b/runs/baseline/checkpoint-3240/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..854d237a4ff7656e8561049ef131e53b73d26e13
--- /dev/null
+++ b/runs/baseline/checkpoint-3240/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:be1f0fdefddf6f30886458ed628f8251f23fec39a294e6649a7ab241784b8460
+size 14917
diff --git a/runs/baseline/checkpoint-3240/rng_state_1.pth b/runs/baseline/checkpoint-3240/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..bf9f97063d7490c565f537598baa2da46c4b6b26
--- /dev/null
+++ b/runs/baseline/checkpoint-3240/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:35669f76342834bd66ba665f73a6cfdbe07c4f2c0f41f6e5e6ed2ad3ae67a161
+size 14917
diff --git a/runs/baseline/checkpoint-3240/scheduler.pt b/runs/baseline/checkpoint-3240/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..1b3635513c9f21d830135bfbbf11f168f3565522
--- /dev/null
+++ b/runs/baseline/checkpoint-3240/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:778f99daa1d0c600722ad8d85ac0f8172450ce32206d332c7edef1f0121201c6
+size 1465
diff --git a/runs/baseline/checkpoint-3240/tokenizer.json b/runs/baseline/checkpoint-3240/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/baseline/checkpoint-3240/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/baseline/checkpoint-3240/tokenizer_config.json b/runs/baseline/checkpoint-3240/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/baseline/checkpoint-3240/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/baseline/checkpoint-3240/trainer_state.json b/runs/baseline/checkpoint-3240/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..b338983818642f2c7c46b66ef54400bf6c2c9f96
--- /dev/null
+++ b/runs/baseline/checkpoint-3240/trainer_state.json
@@ -0,0 +1,23066 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 45.0,
+ "eval_steps": 500,
+ "global_step": 3240,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.12165174633264542,
+ "learning_rate": 0.0,
+ "loss": 12.027063369750977,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.12109383195638657,
+ "learning_rate": 3e-05,
+ "loss": 12.028635025024414,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.12690064311027527,
+ "learning_rate": 6e-05,
+ "loss": 11.894524574279785,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.1537286937236786,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.672262191772461,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.11778244376182556,
+ "learning_rate": 0.00012,
+ "loss": 11.433671951293945,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.10874935984611511,
+ "learning_rate": 0.00015000000000000001,
+ "loss": 11.270309448242188,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.10641921311616898,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 11.155481338500977,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.1048506498336792,
+ "learning_rate": 0.00021000000000000004,
+ "loss": 11.066675186157227,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.10546834021806717,
+ "learning_rate": 0.00024,
+ "loss": 10.952295303344727,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.10614268481731415,
+ "learning_rate": 0.00027,
+ "loss": 10.811386108398438,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10529010742902756,
+ "learning_rate": 0.00030000000000000003,
+ "loss": 10.657712936401367,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.1051444336771965,
+ "learning_rate": 0.00033,
+ "loss": 10.47591781616211,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10420846194028854,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 10.280141830444336,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10323625057935715,
+ "learning_rate": 0.00039000000000000005,
+ "loss": 10.076366424560547,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10255401581525803,
+ "learning_rate": 0.00042000000000000007,
+ "loss": 9.856939315795898,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10151027888059616,
+ "learning_rate": 0.00045,
+ "loss": 9.632134437561035,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.1006326898932457,
+ "learning_rate": 0.00048,
+ "loss": 9.401637077331543,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.09820562601089478,
+ "learning_rate": 0.00051,
+ "loss": 9.177400588989258,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.0956626906991005,
+ "learning_rate": 0.00054,
+ "loss": 8.95518684387207,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.0922091156244278,
+ "learning_rate": 0.00057,
+ "loss": 8.75066089630127,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.08593932539224625,
+ "learning_rate": 0.0006000000000000001,
+ "loss": 8.557976722717285,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.08014766871929169,
+ "learning_rate": 0.00063,
+ "loss": 8.360604286193848,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.07381244748830795,
+ "learning_rate": 0.00066,
+ "loss": 8.210375785827637,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.06455052644014359,
+ "learning_rate": 0.0006900000000000001,
+ "loss": 8.07168960571289,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.054115474224090576,
+ "learning_rate": 0.0007199999999999999,
+ "loss": 7.960069179534912,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.043575264513492584,
+ "learning_rate": 0.00075,
+ "loss": 7.850685119628906,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.02976679615676403,
+ "learning_rate": 0.0007800000000000001,
+ "loss": 7.824377059936523,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.020552165806293488,
+ "learning_rate": 0.0008100000000000001,
+ "loss": 7.7695417404174805,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.018530743196606636,
+ "learning_rate": 0.0008400000000000001,
+ "loss": 7.782247543334961,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.02716820128262043,
+ "learning_rate": 0.00087,
+ "loss": 7.754945278167725,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.03638409823179245,
+ "learning_rate": 0.0009,
+ "loss": 7.799381256103516,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.041579652577638626,
+ "learning_rate": 0.00093,
+ "loss": 7.776629447937012,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.0304960235953331,
+ "learning_rate": 0.00096,
+ "loss": 7.749439239501953,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.053728822618722916,
+ "learning_rate": 0.00099,
+ "loss": 7.745059490203857,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.036322493106126785,
+ "learning_rate": 0.00102,
+ "loss": 7.752074718475342,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.03232959285378456,
+ "learning_rate": 0.00105,
+ "loss": 7.704257011413574,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.03246184438467026,
+ "learning_rate": 0.00108,
+ "loss": 7.667045593261719,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.036655545234680176,
+ "learning_rate": 0.00111,
+ "loss": 7.623445987701416,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.021042725071310997,
+ "learning_rate": 0.00114,
+ "loss": 7.582821846008301,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.022982975468039513,
+ "learning_rate": 0.00117,
+ "loss": 7.58036994934082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.024781864136457443,
+ "learning_rate": 0.0012000000000000001,
+ "loss": 7.5903000831604,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.03667089343070984,
+ "learning_rate": 0.00123,
+ "loss": 7.588233947753906,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.04083017632365227,
+ "learning_rate": 0.00126,
+ "loss": 7.548343658447266,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.02585385926067829,
+ "learning_rate": 0.00129,
+ "loss": 7.604833602905273,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.018793931230902672,
+ "learning_rate": 0.00132,
+ "loss": 7.56705379486084,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.025538936257362366,
+ "learning_rate": 0.00135,
+ "loss": 7.562107086181641,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.02819318138062954,
+ "learning_rate": 0.0013800000000000002,
+ "loss": 7.517213821411133,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.016239255666732788,
+ "learning_rate": 0.00141,
+ "loss": 7.4780755043029785,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.024470701813697815,
+ "learning_rate": 0.0014399999999999999,
+ "loss": 7.505887985229492,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.015499847941100597,
+ "learning_rate": 0.00147,
+ "loss": 7.484469890594482,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.017785200849175453,
+ "learning_rate": 0.0015,
+ "loss": 7.481822967529297,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.015952805057168007,
+ "learning_rate": 0.0015300000000000001,
+ "loss": 7.484804153442383,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.015771260485053062,
+ "learning_rate": 0.0015600000000000002,
+ "loss": 7.4495439529418945,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.024370530620217323,
+ "learning_rate": 0.00159,
+ "loss": 7.471208572387695,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.018173737451434135,
+ "learning_rate": 0.0016200000000000001,
+ "loss": 7.438447952270508,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.020564094185829163,
+ "learning_rate": 0.0016500000000000002,
+ "loss": 7.437922954559326,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.016154944896697998,
+ "learning_rate": 0.0016800000000000003,
+ "loss": 7.429378032684326,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.016817007213830948,
+ "learning_rate": 0.00171,
+ "loss": 7.406428337097168,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.03460947796702385,
+ "learning_rate": 0.00174,
+ "loss": 7.393331527709961,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.15410488843917847,
+ "learning_rate": 0.0017699999999999999,
+ "loss": 7.645856857299805,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.07443148642778397,
+ "learning_rate": 0.0018,
+ "loss": 7.549508094787598,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.0507572740316391,
+ "learning_rate": 0.00183,
+ "loss": 7.517921447753906,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.09682746231555939,
+ "learning_rate": 0.00186,
+ "loss": 7.543978691101074,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.2571517527103424,
+ "learning_rate": 0.00189,
+ "loss": 7.768949508666992,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.0317784883081913,
+ "learning_rate": 0.00192,
+ "loss": 7.509063720703125,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.0353970006108284,
+ "learning_rate": 0.0019500000000000001,
+ "loss": 7.522322654724121,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.03663994371891022,
+ "learning_rate": 0.00198,
+ "loss": 7.504723072052002,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.025481393560767174,
+ "learning_rate": 0.00201,
+ "loss": 7.452597141265869,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.03099915198981762,
+ "learning_rate": 0.00204,
+ "loss": 7.455441474914551,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.022921698167920113,
+ "learning_rate": 0.00207,
+ "loss": 7.4454827308654785,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.032388731837272644,
+ "learning_rate": 0.0021,
+ "loss": 7.467103958129883,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.022428609430789948,
+ "learning_rate": 0.00213,
+ "loss": 7.467540264129639,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.469719409942627,
+ "eval_runtime": 50.9714,
+ "eval_samples_per_second": 47.909,
+ "eval_steps_per_second": 1.511,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.017480911687016487,
+ "learning_rate": 0.00216,
+ "loss": 7.455810546875,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.016675131395459175,
+ "learning_rate": 0.00219,
+ "loss": 7.474848747253418,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.017450228333473206,
+ "learning_rate": 0.00222,
+ "loss": 7.442611217498779,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.016823675483465195,
+ "learning_rate": 0.0022500000000000003,
+ "loss": 7.454628944396973,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.016717543825507164,
+ "learning_rate": 0.00228,
+ "loss": 7.425317764282227,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.01644429750740528,
+ "learning_rate": 0.00231,
+ "loss": 7.4102864265441895,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.013104400597512722,
+ "learning_rate": 0.00234,
+ "loss": 7.4559173583984375,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.01175684668123722,
+ "learning_rate": 0.00237,
+ "loss": 7.438898086547852,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.013736684806644917,
+ "learning_rate": 0.0024000000000000002,
+ "loss": 7.449005603790283,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01176756713539362,
+ "learning_rate": 0.0024300000000000003,
+ "loss": 7.428421974182129,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.011860918253660202,
+ "learning_rate": 0.00246,
+ "loss": 7.411491394042969,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.010093477554619312,
+ "learning_rate": 0.00249,
+ "loss": 7.39311408996582,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.009453288279473782,
+ "learning_rate": 0.00252,
+ "loss": 7.387656211853027,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.010311692021787167,
+ "learning_rate": 0.00255,
+ "loss": 7.401076316833496,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.010053195990622044,
+ "learning_rate": 0.00258,
+ "loss": 7.39655876159668,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.009758037514984608,
+ "learning_rate": 0.00261,
+ "loss": 7.423122882843018,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008394946344196796,
+ "learning_rate": 0.00264,
+ "loss": 7.383601188659668,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.008457187563180923,
+ "learning_rate": 0.00267,
+ "loss": 7.400934219360352,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.010601334273815155,
+ "learning_rate": 0.0027,
+ "loss": 7.400241851806641,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.009009373374283314,
+ "learning_rate": 0.0027300000000000002,
+ "loss": 7.377348899841309,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.00737508200109005,
+ "learning_rate": 0.0027600000000000003,
+ "loss": 7.385881423950195,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.0075711836107075214,
+ "learning_rate": 0.0027900000000000004,
+ "loss": 7.348570823669434,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.00846901722252369,
+ "learning_rate": 0.00282,
+ "loss": 7.364287853240967,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.007160635199397802,
+ "learning_rate": 0.00285,
+ "loss": 7.3881635665893555,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.00682067358866334,
+ "learning_rate": 0.0028799999999999997,
+ "loss": 7.348663330078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.009108101949095726,
+ "learning_rate": 0.00291,
+ "loss": 7.356302261352539,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.015828927978873253,
+ "learning_rate": 0.00294,
+ "loss": 7.347576141357422,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.009683096781373024,
+ "learning_rate": 0.00297,
+ "loss": 7.323967933654785,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.010833270847797394,
+ "learning_rate": 0.003,
+ "loss": 7.3139495849609375,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.014905400574207306,
+ "learning_rate": 0.002999999401781161,
+ "loss": 7.302030086517334,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.023674853146076202,
+ "learning_rate": 0.002999997607125127,
+ "loss": 7.311811923980713,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.050460800528526306,
+ "learning_rate": 0.002999994616033342,
+ "loss": 7.41972541809082,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.02193174883723259,
+ "learning_rate": 0.002999990428508218,
+ "loss": 7.294468879699707,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.036261629313230515,
+ "learning_rate": 0.0029999850445531283,
+ "loss": 7.295568466186523,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.02734929881989956,
+ "learning_rate": 0.002999978464172409,
+ "loss": 7.309239864349365,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.017884107306599617,
+ "learning_rate": 0.002999970687371363,
+ "loss": 7.2427592277526855,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01700308918952942,
+ "learning_rate": 0.0029999617141562573,
+ "loss": 7.241730213165283,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.0209935512393713,
+ "learning_rate": 0.0029999515445343194,
+ "loss": 7.262085437774658,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.023207105696201324,
+ "learning_rate": 0.0029999401785137437,
+ "loss": 7.25450325012207,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.01907353848218918,
+ "learning_rate": 0.002999927616103687,
+ "loss": 7.23149299621582,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.016149666160345078,
+ "learning_rate": 0.0029999138573142715,
+ "loss": 7.210232734680176,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.017513928934931755,
+ "learning_rate": 0.002999898902156582,
+ "loss": 7.218451976776123,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.015002547763288021,
+ "learning_rate": 0.002999882750642667,
+ "loss": 7.169316291809082,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.014233081601560116,
+ "learning_rate": 0.0029998654027855406,
+ "loss": 7.194825172424316,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.01283260714262724,
+ "learning_rate": 0.0029998468585991783,
+ "loss": 7.15669059753418,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.014416799880564213,
+ "learning_rate": 0.0029998271180985223,
+ "loss": 7.143369674682617,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.014337058179080486,
+ "learning_rate": 0.0029998061812994757,
+ "loss": 7.127734184265137,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.016406210139393806,
+ "learning_rate": 0.002999784048218908,
+ "loss": 7.086504936218262,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.01680421642959118,
+ "learning_rate": 0.002999760718874652,
+ "loss": 7.089420318603516,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.010665548034012318,
+ "learning_rate": 0.0029997361932855016,
+ "loss": 7.102463722229004,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013479137793183327,
+ "learning_rate": 0.002999710471471218,
+ "loss": 7.079098701477051,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.012408486567437649,
+ "learning_rate": 0.0029996835534525246,
+ "loss": 7.0790114402771,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.012035196647047997,
+ "learning_rate": 0.0029996554392511094,
+ "loss": 7.058756351470947,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.00899985246360302,
+ "learning_rate": 0.0029996261288896225,
+ "loss": 7.0739240646362305,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.01072906143963337,
+ "learning_rate": 0.002999595622391679,
+ "loss": 7.085537910461426,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.01047342736274004,
+ "learning_rate": 0.0029995639197818577,
+ "loss": 7.060049057006836,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.02342618815600872,
+ "learning_rate": 0.0029995310210857006,
+ "loss": 7.039482116699219,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.011256859637796879,
+ "learning_rate": 0.0029994969263297134,
+ "loss": 7.040412902832031,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.008407019078731537,
+ "learning_rate": 0.0029994616355413666,
+ "loss": 7.0364274978637695,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.011006148532032967,
+ "learning_rate": 0.0029994251487490916,
+ "loss": 6.991468906402588,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.009103255346417427,
+ "learning_rate": 0.0029993874659822875,
+ "loss": 7.001059532165527,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.007875788025557995,
+ "learning_rate": 0.0029993485872713125,
+ "loss": 6.9974260330200195,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.007872611284255981,
+ "learning_rate": 0.0029993085126474913,
+ "loss": 6.980216979980469,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.010071957483887672,
+ "learning_rate": 0.0029992672421431115,
+ "loss": 6.97799015045166,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.01209618803113699,
+ "learning_rate": 0.0029992247757914237,
+ "loss": 6.97202205657959,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.014884858392179012,
+ "learning_rate": 0.0029991811136266425,
+ "loss": 6.976570129394531,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.014953463338315487,
+ "learning_rate": 0.0029991362556839453,
+ "loss": 6.971050262451172,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.008371063508093357,
+ "learning_rate": 0.002999090201999474,
+ "loss": 6.955556869506836,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.011926881037652493,
+ "learning_rate": 0.002999042952610332,
+ "loss": 6.943947792053223,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.013422893360257149,
+ "learning_rate": 0.002998994507554589,
+ "loss": 6.910022735595703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.009672224521636963,
+ "learning_rate": 0.002998944866871275,
+ "loss": 6.8994855880737305,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.012590581551194191,
+ "learning_rate": 0.0029988940306003853,
+ "loss": 6.88709020614624,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.945146083831787,
+ "eval_runtime": 51.5801,
+ "eval_samples_per_second": 47.344,
+ "eval_steps_per_second": 1.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.016554836183786392,
+ "learning_rate": 0.002998841998782877,
+ "loss": 6.916609764099121,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.011085660196840763,
+ "learning_rate": 0.0029987887714606723,
+ "loss": 6.884910583496094,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.008596168830990791,
+ "learning_rate": 0.0029987343486766547,
+ "loss": 6.900066375732422,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.014487302862107754,
+ "learning_rate": 0.0029986787304746715,
+ "loss": 6.8693342208862305,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.015636997297406197,
+ "learning_rate": 0.002998621916899534,
+ "loss": 6.903670787811279,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.013275462202727795,
+ "learning_rate": 0.0029985639079970153,
+ "loss": 6.889361381530762,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.009044744074344635,
+ "learning_rate": 0.002998504703813852,
+ "loss": 6.8549981117248535,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.011327696964144707,
+ "learning_rate": 0.002998444304397745,
+ "loss": 6.856151103973389,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.01310331467539072,
+ "learning_rate": 0.0029983827097973554,
+ "loss": 6.834944725036621,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.010003753937780857,
+ "learning_rate": 0.00299831992006231,
+ "loss": 6.825613975524902,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.009928115643560886,
+ "learning_rate": 0.0029982559352431967,
+ "loss": 6.828935146331787,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.012302150949835777,
+ "learning_rate": 0.0029981907553915674,
+ "loss": 6.821207046508789,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.020555168390274048,
+ "learning_rate": 0.002998124380559936,
+ "loss": 6.8184895515441895,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.011417638510465622,
+ "learning_rate": 0.0029980568108017797,
+ "loss": 6.788754463195801,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.010772978886961937,
+ "learning_rate": 0.002997988046171538,
+ "loss": 6.7861433029174805,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.012753859162330627,
+ "learning_rate": 0.002997918086724614,
+ "loss": 6.788776397705078,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.011327114887535572,
+ "learning_rate": 0.0029978469325173717,
+ "loss": 6.779160022735596,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01235753484070301,
+ "learning_rate": 0.002997774583607139,
+ "loss": 6.799554824829102,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.01417156495153904,
+ "learning_rate": 0.0029977010400522074,
+ "loss": 6.77957820892334,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.021321553736925125,
+ "learning_rate": 0.002997626301911828,
+ "loss": 6.785632133483887,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.03614472970366478,
+ "learning_rate": 0.002997550369246217,
+ "loss": 6.836414337158203,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03934052586555481,
+ "learning_rate": 0.002997473242116551,
+ "loss": 6.854659557342529,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.016240060329437256,
+ "learning_rate": 0.0029973949205849707,
+ "loss": 6.789690971374512,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.031164905056357384,
+ "learning_rate": 0.0029973154047145782,
+ "loss": 6.8363776206970215,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.013051844201982021,
+ "learning_rate": 0.002997234694569438,
+ "loss": 6.803576469421387,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.022611448541283607,
+ "learning_rate": 0.002997152790214577,
+ "loss": 6.762107849121094,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.017812907695770264,
+ "learning_rate": 0.0029970696917159833,
+ "loss": 6.776218414306641,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.015453332103788853,
+ "learning_rate": 0.0029969853991406088,
+ "loss": 6.742618083953857,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.020630091428756714,
+ "learning_rate": 0.0029968999125563655,
+ "loss": 6.765475749969482,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.014672341756522655,
+ "learning_rate": 0.0029968132320321297,
+ "loss": 6.754009246826172,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.016540220007300377,
+ "learning_rate": 0.0029967253576377376,
+ "loss": 6.734752655029297,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.013495649211108685,
+ "learning_rate": 0.0029966362894439874,
+ "loss": 6.752676010131836,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015936989337205887,
+ "learning_rate": 0.0029965460275226405,
+ "loss": 6.725708961486816,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010431243106722832,
+ "learning_rate": 0.0029964545719464194,
+ "loss": 6.698376655578613,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01669321395456791,
+ "learning_rate": 0.0029963619227890083,
+ "loss": 6.711770534515381,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.013308997265994549,
+ "learning_rate": 0.0029962680801250514,
+ "loss": 6.705286979675293,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011648476123809814,
+ "learning_rate": 0.002996173044030159,
+ "loss": 6.690052509307861,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.03942161798477173,
+ "learning_rate": 0.0029960768145808965,
+ "loss": 6.711885452270508,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.01571274921298027,
+ "learning_rate": 0.002995979391854797,
+ "loss": 6.715847969055176,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.00829714722931385,
+ "learning_rate": 0.0029958807759303515,
+ "loss": 6.685308933258057,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.01374727487564087,
+ "learning_rate": 0.0029957809668870114,
+ "loss": 6.695744514465332,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.010857059620320797,
+ "learning_rate": 0.0029956799648051934,
+ "loss": 6.658294677734375,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.008897246792912483,
+ "learning_rate": 0.002995577769766272,
+ "loss": 6.642439842224121,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015880772843956947,
+ "learning_rate": 0.002995474381852584,
+ "loss": 6.674907684326172,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.021623898297548294,
+ "learning_rate": 0.0029953698011474265,
+ "loss": 6.66236686706543,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.03084627166390419,
+ "learning_rate": 0.0029952640277350596,
+ "loss": 6.666715621948242,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.040561504662036896,
+ "learning_rate": 0.002995157061700702,
+ "loss": 6.649062156677246,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.031126515939831734,
+ "learning_rate": 0.0029950489031305345,
+ "loss": 6.640542030334473,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.00872360821813345,
+ "learning_rate": 0.0029949395521116987,
+ "loss": 6.635221481323242,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.023432321846485138,
+ "learning_rate": 0.0029948290087322966,
+ "loss": 6.642196178436279,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.021096013486385345,
+ "learning_rate": 0.0029947172730813913,
+ "loss": 6.643655300140381,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.014494858682155609,
+ "learning_rate": 0.002994604345249006,
+ "loss": 6.612720966339111,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.05225599184632301,
+ "learning_rate": 0.0029944902253261248,
+ "loss": 6.60659646987915,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.014311269856989384,
+ "learning_rate": 0.0029943749134046915,
+ "loss": 6.54373025894165,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.010777472518384457,
+ "learning_rate": 0.0029942584095776115,
+ "loss": 6.599837303161621,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.012498079799115658,
+ "learning_rate": 0.002994140713938749,
+ "loss": 6.589657783508301,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.015775132924318314,
+ "learning_rate": 0.00299402182658293,
+ "loss": 6.577511787414551,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.017978230491280556,
+ "learning_rate": 0.0029939017476059396,
+ "loss": 6.58717155456543,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.026807745918631554,
+ "learning_rate": 0.0029937804771045236,
+ "loss": 6.578145503997803,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.04853282868862152,
+ "learning_rate": 0.0029936580151763872,
+ "loss": 6.609755516052246,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.1057916209101677,
+ "learning_rate": 0.002993534361920195,
+ "loss": 6.870367050170898,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.04925911873579025,
+ "learning_rate": 0.002993409517435574,
+ "loss": 6.871556282043457,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.05288556218147278,
+ "learning_rate": 0.0029932834818231077,
+ "loss": 6.808315753936768,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.03831710293889046,
+ "learning_rate": 0.0029931562551843405,
+ "loss": 6.827617645263672,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.03665117174386978,
+ "learning_rate": 0.002993027837621778,
+ "loss": 6.767284870147705,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.026216736063361168,
+ "learning_rate": 0.0029928982292388826,
+ "loss": 6.733659744262695,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.024796854704618454,
+ "learning_rate": 0.002992767430140078,
+ "loss": 6.759577751159668,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.02410992607474327,
+ "learning_rate": 0.002992635440430747,
+ "loss": 6.701825141906738,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.019924134016036987,
+ "learning_rate": 0.0029925022602172308,
+ "loss": 6.677121162414551,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.022516867145895958,
+ "learning_rate": 0.00299236788960683,
+ "loss": 6.668622970581055,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.025068556889891624,
+ "learning_rate": 0.002992232328707806,
+ "loss": 6.67718505859375,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.02574208751320839,
+ "learning_rate": 0.002992095577629376,
+ "loss": 6.679594039916992,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.657555103302002,
+ "eval_runtime": 51.0199,
+ "eval_samples_per_second": 47.864,
+ "eval_steps_per_second": 1.509,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.017298126593232155,
+ "learning_rate": 0.0029919576364817184,
+ "loss": 6.620774269104004,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.018840325996279716,
+ "learning_rate": 0.002991818505375971,
+ "loss": 6.613316535949707,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.020813113078475,
+ "learning_rate": 0.002991678184424228,
+ "loss": 6.621555328369141,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.026581479236483574,
+ "learning_rate": 0.0029915366737395437,
+ "loss": 6.613471031188965,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.02084030583500862,
+ "learning_rate": 0.0029913939734359306,
+ "loss": 6.559704780578613,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.01327502354979515,
+ "learning_rate": 0.00299125008362836,
+ "loss": 6.564910888671875,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.015091734007000923,
+ "learning_rate": 0.0029911050044327614,
+ "loss": 6.571262836456299,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.015463982708752155,
+ "learning_rate": 0.002990958735966023,
+ "loss": 6.565750598907471,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.013636826537549496,
+ "learning_rate": 0.0029908112783459894,
+ "loss": 6.57981014251709,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.010613273829221725,
+ "learning_rate": 0.0029906626316914655,
+ "loss": 6.538505554199219,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.01804237812757492,
+ "learning_rate": 0.0029905127961222127,
+ "loss": 6.5282511711120605,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.015131509862840176,
+ "learning_rate": 0.0029903617717589506,
+ "loss": 6.531335830688477,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.010518206283450127,
+ "learning_rate": 0.002990209558723358,
+ "loss": 6.521886825561523,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.011360022239387035,
+ "learning_rate": 0.002990056157138069,
+ "loss": 6.528301239013672,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.011556181125342846,
+ "learning_rate": 0.0029899015671266773,
+ "loss": 6.497917652130127,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.01365791168063879,
+ "learning_rate": 0.0029897457888137325,
+ "loss": 6.52849006652832,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.01091043371707201,
+ "learning_rate": 0.0029895888223247426,
+ "loss": 6.492197036743164,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.008484847843647003,
+ "learning_rate": 0.002989430667786173,
+ "loss": 6.497377395629883,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.00836879201233387,
+ "learning_rate": 0.0029892713253254466,
+ "loss": 6.495189666748047,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.009781497530639172,
+ "learning_rate": 0.0029891107950709406,
+ "loss": 6.454085826873779,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.009603648446500301,
+ "learning_rate": 0.0029889490771519934,
+ "loss": 6.499475479125977,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.007991855964064598,
+ "learning_rate": 0.0029887861716988975,
+ "loss": 6.460785865783691,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.008447187952697277,
+ "learning_rate": 0.0029886220788429025,
+ "loss": 6.504582405090332,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.010945498012006283,
+ "learning_rate": 0.0029884567987162156,
+ "loss": 6.49589729309082,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.011096215806901455,
+ "learning_rate": 0.0029882903314519997,
+ "loss": 6.460657119750977,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.013094739057123661,
+ "learning_rate": 0.0029881226771843747,
+ "loss": 6.462450981140137,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.015423037111759186,
+ "learning_rate": 0.0029879538360484165,
+ "loss": 6.45565128326416,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.0182222630828619,
+ "learning_rate": 0.002987783808180157,
+ "loss": 6.452227592468262,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.01981227844953537,
+ "learning_rate": 0.002987612593716585,
+ "loss": 6.477701663970947,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.026733700186014175,
+ "learning_rate": 0.002987440192795645,
+ "loss": 6.452212333679199,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.03504155948758125,
+ "learning_rate": 0.0029872666055562372,
+ "loss": 6.509665489196777,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.027970125898718834,
+ "learning_rate": 0.002987091832138217,
+ "loss": 6.487089157104492,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.013714591972529888,
+ "learning_rate": 0.0029869158726823974,
+ "loss": 6.469569206237793,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.016472695395350456,
+ "learning_rate": 0.0029867387273305453,
+ "loss": 6.461688995361328,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.018457358703017235,
+ "learning_rate": 0.0029865603962253833,
+ "loss": 6.490985870361328,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.011045100167393684,
+ "learning_rate": 0.0029863808795105896,
+ "loss": 6.403701305389404,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.014126897789537907,
+ "learning_rate": 0.0029862001773307972,
+ "loss": 6.379804611206055,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.014015335589647293,
+ "learning_rate": 0.0029860182898315954,
+ "loss": 6.404377460479736,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.010032658465206623,
+ "learning_rate": 0.002985835217159527,
+ "loss": 6.40291690826416,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.015469838865101337,
+ "learning_rate": 0.0029856509594620912,
+ "loss": 6.4414520263671875,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.011106910184025764,
+ "learning_rate": 0.0029854655168877397,
+ "loss": 6.396444320678711,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.01262320950627327,
+ "learning_rate": 0.0029852788895858816,
+ "loss": 6.385222434997559,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.012783976271748543,
+ "learning_rate": 0.002985091077706878,
+ "loss": 6.3818135261535645,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.01224067248404026,
+ "learning_rate": 0.0029849020814020468,
+ "loss": 6.37886905670166,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.010540086776018143,
+ "learning_rate": 0.0029847119008236576,
+ "loss": 6.37799072265625,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.012479913420975208,
+ "learning_rate": 0.0029845205361249366,
+ "loss": 6.365094184875488,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.010080077685415745,
+ "learning_rate": 0.0029843279874600622,
+ "loss": 6.371852874755859,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.013266298919916153,
+ "learning_rate": 0.002984134254984168,
+ "loss": 6.380039215087891,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.01014771405607462,
+ "learning_rate": 0.002983939338853339,
+ "loss": 6.370337963104248,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.008164579048752785,
+ "learning_rate": 0.0029837432392246185,
+ "loss": 6.331899642944336,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.01020442508161068,
+ "learning_rate": 0.002983545956255998,
+ "loss": 6.35077428817749,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.009898579679429531,
+ "learning_rate": 0.002983347490106426,
+ "loss": 6.371985912322998,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.009668919257819653,
+ "learning_rate": 0.002983147840935802,
+ "loss": 6.356776237487793,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.012560366652905941,
+ "learning_rate": 0.0029829470089049815,
+ "loss": 6.357911109924316,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.018435249105095863,
+ "learning_rate": 0.0029827449941757698,
+ "loss": 6.366821765899658,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.029046855866909027,
+ "learning_rate": 0.002982541796910927,
+ "loss": 6.367595672607422,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.043010637164115906,
+ "learning_rate": 0.002982337417274165,
+ "loss": 6.413588523864746,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.045470818877220154,
+ "learning_rate": 0.0029821318554301497,
+ "loss": 6.435015678405762,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.02148726023733616,
+ "learning_rate": 0.002981925111544498,
+ "loss": 6.330894470214844,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.026808688417077065,
+ "learning_rate": 0.00298171718578378,
+ "loss": 6.384839057922363,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.018495213240385056,
+ "learning_rate": 0.002981508078315517,
+ "loss": 6.36277961730957,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.025393597781658173,
+ "learning_rate": 0.0029812977893081834,
+ "loss": 6.3702239990234375,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020856289193034172,
+ "learning_rate": 0.002981086318931206,
+ "loss": 6.354547500610352,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.017827816307544708,
+ "learning_rate": 0.002980873667354962,
+ "loss": 6.339264869689941,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.02639714442193508,
+ "learning_rate": 0.002980659834750781,
+ "loss": 6.3254194259643555,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.015238096937537193,
+ "learning_rate": 0.002980444821290944,
+ "loss": 6.365842819213867,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.012863729149103165,
+ "learning_rate": 0.002980228627148684,
+ "loss": 6.356776714324951,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.014739899896085262,
+ "learning_rate": 0.0029800112524981845,
+ "loss": 6.332481384277344,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.013582171872258186,
+ "learning_rate": 0.0029797926975145797,
+ "loss": 6.311968803405762,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.012704637832939625,
+ "learning_rate": 0.0029795729623739555,
+ "loss": 6.321654319763184,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.012550560757517815,
+ "learning_rate": 0.0029793520472533495,
+ "loss": 6.298409461975098,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.011170165613293648,
+ "learning_rate": 0.002979129952330748,
+ "loss": 6.294830799102783,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.326757907867432,
+ "eval_runtime": 51.6566,
+ "eval_samples_per_second": 47.274,
+ "eval_steps_per_second": 1.491,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.010947446338832378,
+ "learning_rate": 0.0029789066777850897,
+ "loss": 6.286581039428711,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.010097995400428772,
+ "learning_rate": 0.002978682223796262,
+ "loss": 6.288301467895508,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.01224475260823965,
+ "learning_rate": 0.002978456590545104,
+ "loss": 6.264986991882324,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.009734537452459335,
+ "learning_rate": 0.0029782297782134045,
+ "loss": 6.304243087768555,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.01136956550180912,
+ "learning_rate": 0.0029780017869839013,
+ "loss": 6.244938373565674,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.00935293547809124,
+ "learning_rate": 0.002977772617040284,
+ "loss": 6.251852035522461,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.011118784546852112,
+ "learning_rate": 0.0029775422685671894,
+ "loss": 6.256633758544922,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.010837923735380173,
+ "learning_rate": 0.0029773107417502055,
+ "loss": 6.263411045074463,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.011920305900275707,
+ "learning_rate": 0.0029770780367758703,
+ "loss": 6.239311218261719,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.014117483049631119,
+ "learning_rate": 0.0029768441538316685,
+ "loss": 6.253314018249512,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.022502148523926735,
+ "learning_rate": 0.002976609093106036,
+ "loss": 6.246156692504883,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.03348720073699951,
+ "learning_rate": 0.002976372854788357,
+ "loss": 6.255914688110352,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.03427121415734291,
+ "learning_rate": 0.002976135439068965,
+ "loss": 6.289946556091309,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.02407355234026909,
+ "learning_rate": 0.002975896846139141,
+ "loss": 6.288534164428711,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.017861567437648773,
+ "learning_rate": 0.0029756570761911147,
+ "loss": 6.258049964904785,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.019980227574706078,
+ "learning_rate": 0.0029754161294180646,
+ "loss": 6.250895023345947,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.01443127728998661,
+ "learning_rate": 0.0029751740060141177,
+ "loss": 6.212200164794922,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.014975055120885372,
+ "learning_rate": 0.0029749307061743475,
+ "loss": 6.230568885803223,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01408296450972557,
+ "learning_rate": 0.0029746862300947773,
+ "loss": 6.243950366973877,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.011649327352643013,
+ "learning_rate": 0.0029744405779723766,
+ "loss": 6.197744846343994,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.013392188586294651,
+ "learning_rate": 0.002974193750005063,
+ "loss": 6.207277297973633,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.010902997106313705,
+ "learning_rate": 0.002973945746391701,
+ "loss": 6.166373252868652,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.012915392406284809,
+ "learning_rate": 0.0029736965673321032,
+ "loss": 6.186316013336182,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.010772830806672573,
+ "learning_rate": 0.0029734462130270273,
+ "loss": 6.172578811645508,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.010274911299347878,
+ "learning_rate": 0.002973194683678181,
+ "loss": 6.206851005554199,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.013210494071245193,
+ "learning_rate": 0.0029729419794882164,
+ "loss": 6.21115779876709,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.013372313231229782,
+ "learning_rate": 0.0029726881006607326,
+ "loss": 6.18442440032959,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.008105908520519733,
+ "learning_rate": 0.0029724330474002743,
+ "loss": 6.204354286193848,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.014167108573019505,
+ "learning_rate": 0.0029721768199123346,
+ "loss": 6.180662155151367,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.008339803665876389,
+ "learning_rate": 0.002971919418403351,
+ "loss": 6.150079727172852,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.015939386561512947,
+ "learning_rate": 0.0029716608430807073,
+ "loss": 6.175203323364258,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.010777109302580357,
+ "learning_rate": 0.0029714010941527327,
+ "loss": 6.166106224060059,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.011274735443294048,
+ "learning_rate": 0.002971140171828702,
+ "loss": 6.163111686706543,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013626808300614357,
+ "learning_rate": 0.002970878076318836,
+ "loss": 6.140850067138672,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.016907762736082077,
+ "learning_rate": 0.0029706148078343,
+ "loss": 6.192012786865234,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.018817441537976265,
+ "learning_rate": 0.002970350366587205,
+ "loss": 6.162238121032715,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.02829768694937229,
+ "learning_rate": 0.0029700847527906068,
+ "loss": 6.159066200256348,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.02034052088856697,
+ "learning_rate": 0.002969817966658505,
+ "loss": 6.164158344268799,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.01816369965672493,
+ "learning_rate": 0.0029695500084058447,
+ "loss": 6.132316589355469,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.013685261830687523,
+ "learning_rate": 0.0029692808782485146,
+ "loss": 6.152181625366211,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.009791149757802486,
+ "learning_rate": 0.0029690105764033485,
+ "loss": 6.138544082641602,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.014827069826424122,
+ "learning_rate": 0.002968739103088124,
+ "loss": 6.135766983032227,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.01391249243170023,
+ "learning_rate": 0.0029684664585215618,
+ "loss": 6.146172523498535,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.012729899026453495,
+ "learning_rate": 0.002968192642923326,
+ "loss": 6.122613430023193,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.014175458811223507,
+ "learning_rate": 0.002967917656514026,
+ "loss": 6.12574577331543,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.010958840139210224,
+ "learning_rate": 0.002967641499515213,
+ "loss": 6.111088752746582,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.015380816534161568,
+ "learning_rate": 0.0029673641721493817,
+ "loss": 6.100615501403809,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.025339096784591675,
+ "learning_rate": 0.00296708567463997,
+ "loss": 6.139659881591797,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.03695995733141899,
+ "learning_rate": 0.002966806007211358,
+ "loss": 6.160996437072754,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.0408046580851078,
+ "learning_rate": 0.0029665251700888686,
+ "loss": 6.173980712890625,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.025022024288773537,
+ "learning_rate": 0.0029662431634987674,
+ "loss": 6.145794868469238,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.027254154905676842,
+ "learning_rate": 0.002965959987668262,
+ "loss": 6.16928243637085,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.06191903352737427,
+ "learning_rate": 0.0029656756428255017,
+ "loss": 6.138525009155273,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0348956398665905,
+ "learning_rate": 0.002965390129199579,
+ "loss": 6.185477256774902,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.020859677344560623,
+ "learning_rate": 0.002965103447020526,
+ "loss": 6.148846626281738,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.022810814902186394,
+ "learning_rate": 0.0029648155965193174,
+ "loss": 6.1701178550720215,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.017981817945837975,
+ "learning_rate": 0.00296452657792787,
+ "loss": 6.1323771476745605,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.014905866235494614,
+ "learning_rate": 0.00296423639147904,
+ "loss": 6.123426914215088,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.01463468000292778,
+ "learning_rate": 0.002963945037406626,
+ "loss": 6.146122932434082,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.021616583690047264,
+ "learning_rate": 0.002963652515945366,
+ "loss": 6.093459606170654,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.01550968922674656,
+ "learning_rate": 0.00296335882733094,
+ "loss": 6.114929676055908,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.014597428031265736,
+ "learning_rate": 0.0029630639717999675,
+ "loss": 6.113945960998535,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.018249280750751495,
+ "learning_rate": 0.002962767949590008,
+ "loss": 6.130507469177246,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.014160148799419403,
+ "learning_rate": 0.0029624707609395614,
+ "loss": 6.086946487426758,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.015563832595944405,
+ "learning_rate": 0.002962172406088067,
+ "loss": 6.083332538604736,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.012240015901625156,
+ "learning_rate": 0.002961872885275904,
+ "loss": 6.0822906494140625,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.0141651201993227,
+ "learning_rate": 0.002961572198744391,
+ "loss": 6.049562454223633,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.010004645213484764,
+ "learning_rate": 0.0029612703467357866,
+ "loss": 6.067063808441162,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.011790835298597813,
+ "learning_rate": 0.0029609673294932856,
+ "loss": 6.062193870544434,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.011088060215115547,
+ "learning_rate": 0.002960663147261025,
+ "loss": 6.070950031280518,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.010387620888650417,
+ "learning_rate": 0.002960357800284078,
+ "loss": 6.032423973083496,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.012074458412826061,
+ "learning_rate": 0.002960051288808458,
+ "loss": 6.066788673400879,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.097111701965332,
+ "eval_runtime": 51.8742,
+ "eval_samples_per_second": 47.075,
+ "eval_steps_per_second": 1.484,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.013429952785372734,
+ "learning_rate": 0.0029597436130811153,
+ "loss": 6.022172451019287,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.01764705777168274,
+ "learning_rate": 0.0029594347733499387,
+ "loss": 6.029745101928711,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.02403099089860916,
+ "learning_rate": 0.002959124769863755,
+ "loss": 6.04707145690918,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.029030559584498405,
+ "learning_rate": 0.002958813602872328,
+ "loss": 6.05479097366333,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.029980644583702087,
+ "learning_rate": 0.0029585012726263595,
+ "loss": 6.068657875061035,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.021424276754260063,
+ "learning_rate": 0.0029581877793774885,
+ "loss": 6.062008857727051,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.018076395615935326,
+ "learning_rate": 0.002957873123378291,
+ "loss": 6.023279190063477,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.017688792198896408,
+ "learning_rate": 0.002957557304882279,
+ "loss": 6.014641284942627,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.02313626930117607,
+ "learning_rate": 0.0029572403241439026,
+ "loss": 6.026420593261719,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.04045062139630318,
+ "learning_rate": 0.0029569221814185467,
+ "loss": 6.007199764251709,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.015792209655046463,
+ "learning_rate": 0.0029566028769625334,
+ "loss": 6.017209053039551,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.01990479789674282,
+ "learning_rate": 0.002956282411033121,
+ "loss": 5.982841491699219,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.01191043108701706,
+ "learning_rate": 0.002955960783888503,
+ "loss": 6.0035929679870605,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.01732831820845604,
+ "learning_rate": 0.002955637995787807,
+ "loss": 6.025006294250488,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.032948266714811325,
+ "learning_rate": 0.0029553140469910995,
+ "loss": 6.007736682891846,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.011192727833986282,
+ "learning_rate": 0.00295498893775938,
+ "loss": 6.0022687911987305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014646840281784534,
+ "learning_rate": 0.0029546626683545824,
+ "loss": 5.976479530334473,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01589796133339405,
+ "learning_rate": 0.002954335239039576,
+ "loss": 5.980949878692627,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.015129734762012959,
+ "learning_rate": 0.0029540066500781657,
+ "loss": 6.0073394775390625,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.011140518821775913,
+ "learning_rate": 0.002953676901735088,
+ "loss": 5.976587295532227,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.0109480582177639,
+ "learning_rate": 0.0029533459942760174,
+ "loss": 6.001427173614502,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.013346790336072445,
+ "learning_rate": 0.0029530139279675583,
+ "loss": 5.96919059753418,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.012474581599235535,
+ "learning_rate": 0.002952680703077251,
+ "loss": 5.970426559448242,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01140566635876894,
+ "learning_rate": 0.002952346319873569,
+ "loss": 5.989171504974365,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.0137029392644763,
+ "learning_rate": 0.0029520107786259187,
+ "loss": 6.010115623474121,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015517761930823326,
+ "learning_rate": 0.002951674079604639,
+ "loss": 6.003649711608887,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.02173629403114319,
+ "learning_rate": 0.0029513362230810035,
+ "loss": 5.967074871063232,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02683914266526699,
+ "learning_rate": 0.002950997209327216,
+ "loss": 5.994434833526611,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.025865858420729637,
+ "learning_rate": 0.002950657038616415,
+ "loss": 6.010523796081543,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02544897235929966,
+ "learning_rate": 0.002950315711222668,
+ "loss": 5.994078159332275,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.028426019474864006,
+ "learning_rate": 0.0029499732274209777,
+ "loss": 6.020925521850586,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.026174750179052353,
+ "learning_rate": 0.0029496295874872768,
+ "loss": 5.985616683959961,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.02802138403058052,
+ "learning_rate": 0.00294928479169843,
+ "loss": 5.973659038543701,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.019348938018083572,
+ "learning_rate": 0.002948938840332233,
+ "loss": 5.990137577056885,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.016294041648507118,
+ "learning_rate": 0.0029485917336674115,
+ "loss": 5.962082862854004,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.016167785972356796,
+ "learning_rate": 0.002948243471983625,
+ "loss": 5.9594573974609375,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.014119232073426247,
+ "learning_rate": 0.00294789405556146,
+ "loss": 5.894467353820801,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.013805347494781017,
+ "learning_rate": 0.002947543484682436,
+ "loss": 5.964014530181885,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.014496497809886932,
+ "learning_rate": 0.002947191759629001,
+ "loss": 5.989920616149902,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.014813800342381,
+ "learning_rate": 0.0029468388806845333,
+ "loss": 5.9669413566589355,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.01153622381389141,
+ "learning_rate": 0.0029464848481333423,
+ "loss": 5.918018341064453,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.010525977239012718,
+ "learning_rate": 0.0029461296622606647,
+ "loss": 5.892477989196777,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.011410463601350784,
+ "learning_rate": 0.002945773323352667,
+ "loss": 5.925249099731445,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.010949424467980862,
+ "learning_rate": 0.0029454158316964464,
+ "loss": 5.90645694732666,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.00962804164737463,
+ "learning_rate": 0.002945057187580026,
+ "loss": 5.917054176330566,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.011517546139657497,
+ "learning_rate": 0.0029446973912923597,
+ "loss": 5.919445514678955,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01567666605114937,
+ "learning_rate": 0.0029443364431233285,
+ "loss": 5.884111404418945,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.016781140118837357,
+ "learning_rate": 0.002943974343363742,
+ "loss": 5.9252119064331055,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.02244546078145504,
+ "learning_rate": 0.0029436110923053375,
+ "loss": 5.933143615722656,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.028186792507767677,
+ "learning_rate": 0.00294324669024078,
+ "loss": 5.9079132080078125,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.028585223481059074,
+ "learning_rate": 0.002942881137463661,
+ "loss": 5.925262451171875,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.02781485952436924,
+ "learning_rate": 0.0029425144342685003,
+ "loss": 5.963315010070801,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.021486928686499596,
+ "learning_rate": 0.002942146580950744,
+ "loss": 5.930115699768066,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.016890818253159523,
+ "learning_rate": 0.002941777577806765,
+ "loss": 5.921439170837402,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.0169126745313406,
+ "learning_rate": 0.0029414074251338618,
+ "loss": 5.928220272064209,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.014659176580607891,
+ "learning_rate": 0.002941036123230261,
+ "loss": 5.934546947479248,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.015337883494794369,
+ "learning_rate": 0.002940663672395112,
+ "loss": 5.890145301818848,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01254988368600607,
+ "learning_rate": 0.002940290072928493,
+ "loss": 5.875492095947266,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.014658362604677677,
+ "learning_rate": 0.0029399153251314066,
+ "loss": 5.901272773742676,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.017342250794172287,
+ "learning_rate": 0.002939539429305779,
+ "loss": 5.875853538513184,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.016825608909130096,
+ "learning_rate": 0.0029391623857544643,
+ "loss": 5.941972732543945,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.016716884449124336,
+ "learning_rate": 0.002938784194781238,
+ "loss": 5.912367343902588,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.021707067266106606,
+ "learning_rate": 0.002938404856690803,
+ "loss": 5.895319938659668,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.030519135296344757,
+ "learning_rate": 0.002938024371788785,
+ "loss": 5.897165298461914,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.02572188340127468,
+ "learning_rate": 0.002937642740381733,
+ "loss": 5.879164695739746,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.016286717727780342,
+ "learning_rate": 0.002937259962777122,
+ "loss": 5.841421127319336,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.020656082779169083,
+ "learning_rate": 0.002936876039283347,
+ "loss": 5.9222259521484375,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.015350459143519402,
+ "learning_rate": 0.0029364909702097298,
+ "loss": 5.885847091674805,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01708570495247841,
+ "learning_rate": 0.0029361047558665136,
+ "loss": 5.872044563293457,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.013537376187741756,
+ "learning_rate": 0.0029357173965648634,
+ "loss": 5.866434097290039,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.013208637945353985,
+ "learning_rate": 0.002935328892616868,
+ "loss": 5.8624491691589355,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.012249554507434368,
+ "learning_rate": 0.002934939244335538,
+ "loss": 5.833232879638672,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.90985107421875,
+ "eval_runtime": 51.0696,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.01382362935692072,
+ "learning_rate": 0.0029345484520348062,
+ "loss": 5.835695266723633,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.012947693467140198,
+ "learning_rate": 0.0029341565160295262,
+ "loss": 5.826190948486328,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.013653872534632683,
+ "learning_rate": 0.002933763436635475,
+ "loss": 5.831789016723633,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.027596980333328247,
+ "learning_rate": 0.0029333692141693484,
+ "loss": 5.839963436126709,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01637990213930607,
+ "learning_rate": 0.0029329738489487648,
+ "loss": 5.805248260498047,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.02169182151556015,
+ "learning_rate": 0.0029325773412922627,
+ "loss": 5.838248252868652,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.03518201410770416,
+ "learning_rate": 0.0029321796915193015,
+ "loss": 5.834336280822754,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04983877018094063,
+ "learning_rate": 0.0029317808999502603,
+ "loss": 5.879856109619141,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05085664615035057,
+ "learning_rate": 0.0029313809669064377,
+ "loss": 5.946531295776367,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.03314858302474022,
+ "learning_rate": 0.0029309798927100532,
+ "loss": 5.909746170043945,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.034118980169296265,
+ "learning_rate": 0.0029305776776842453,
+ "loss": 5.903947830200195,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.03401247411966324,
+ "learning_rate": 0.0029301743221530705,
+ "loss": 5.936162948608398,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.043827857822179794,
+ "learning_rate": 0.0029297698264415047,
+ "loss": 5.953949451446533,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.06281330436468124,
+ "learning_rate": 0.002929364190875444,
+ "loss": 5.988225936889648,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.07216024398803711,
+ "learning_rate": 0.0029289574157817014,
+ "loss": 6.187741279602051,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.055205658078193665,
+ "learning_rate": 0.002928549501488008,
+ "loss": 6.177831649780273,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.04834286496043205,
+ "learning_rate": 0.0029281404483230133,
+ "loss": 6.209604263305664,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.0447777584195137,
+ "learning_rate": 0.0029277302566162837,
+ "loss": 6.146507263183594,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.05729430541396141,
+ "learning_rate": 0.0029273189266983035,
+ "loss": 6.280686378479004,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.0410807766020298,
+ "learning_rate": 0.002926906458900473,
+ "loss": 6.246824264526367,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.044607777148485184,
+ "learning_rate": 0.0029264928535551104,
+ "loss": 6.19956111907959,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.05141043663024902,
+ "learning_rate": 0.0029260781109954515,
+ "loss": 6.118578910827637,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.029248764738440514,
+ "learning_rate": 0.0029256622315556448,
+ "loss": 6.155360221862793,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.028403814882040024,
+ "learning_rate": 0.0029252452155707582,
+ "loss": 6.097352027893066,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.025872666388750076,
+ "learning_rate": 0.002924827063376773,
+ "loss": 6.072485446929932,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.018447555601596832,
+ "learning_rate": 0.0029244077753105874,
+ "loss": 6.077045440673828,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.023166058585047722,
+ "learning_rate": 0.0029239873517100147,
+ "loss": 6.040050983428955,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.019579261541366577,
+ "learning_rate": 0.002923565792913782,
+ "loss": 6.039424896240234,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.018369073048233986,
+ "learning_rate": 0.002923143099261531,
+ "loss": 6.021180152893066,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.014188461937010288,
+ "learning_rate": 0.0029227192710938195,
+ "loss": 5.9668049812316895,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.019567398354411125,
+ "learning_rate": 0.0029222943087521182,
+ "loss": 5.989189147949219,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.014190621674060822,
+ "learning_rate": 0.0029218682125788102,
+ "loss": 5.942726135253906,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01375526562333107,
+ "learning_rate": 0.002921440982917195,
+ "loss": 5.942777633666992,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.014203300699591637,
+ "learning_rate": 0.0029210126201114836,
+ "loss": 5.927928924560547,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.014231069944798946,
+ "learning_rate": 0.0029205831245067994,
+ "loss": 5.923510551452637,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.012771515175700188,
+ "learning_rate": 0.0029201524964491796,
+ "loss": 5.908803939819336,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.013399476185441017,
+ "learning_rate": 0.0029197207362855735,
+ "loss": 5.890079021453857,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.025862423703074455,
+ "learning_rate": 0.0029192878443638427,
+ "loss": 5.919907569885254,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.011315986514091492,
+ "learning_rate": 0.0029188538210327595,
+ "loss": 5.901529788970947,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.01235947571694851,
+ "learning_rate": 0.00291841866664201,
+ "loss": 5.903595924377441,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.012175127863883972,
+ "learning_rate": 0.0029179823815421893,
+ "loss": 5.868306636810303,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01195445004850626,
+ "learning_rate": 0.0029175449660848052,
+ "loss": 5.899898529052734,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.012801481410861015,
+ "learning_rate": 0.0029171064206222745,
+ "loss": 5.890621662139893,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.011909718625247478,
+ "learning_rate": 0.002916666745507926,
+ "loss": 5.86421012878418,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01162486057728529,
+ "learning_rate": 0.0029162259410959977,
+ "loss": 5.869139671325684,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.010923407040536404,
+ "learning_rate": 0.0029157840077416374,
+ "loss": 5.869789123535156,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.011943561024963856,
+ "learning_rate": 0.002915340945800904,
+ "loss": 5.873437881469727,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.010295520536601543,
+ "learning_rate": 0.002914896755630763,
+ "loss": 5.819953918457031,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.0093373479321599,
+ "learning_rate": 0.002914451437589092,
+ "loss": 5.829481601715088,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.01182442344725132,
+ "learning_rate": 0.002914004992034674,
+ "loss": 5.855249404907227,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.00939213763922453,
+ "learning_rate": 0.0029135574193272034,
+ "loss": 5.83876371383667,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.011648556217551231,
+ "learning_rate": 0.0029131087198272813,
+ "loss": 5.824906826019287,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.011227169074118137,
+ "learning_rate": 0.002912658893896416,
+ "loss": 5.847869873046875,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.011084851808845997,
+ "learning_rate": 0.002912207941897025,
+ "loss": 5.839183807373047,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.009923825040459633,
+ "learning_rate": 0.002911755864192432,
+ "loss": 5.818821907043457,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.011300318874418736,
+ "learning_rate": 0.002911302661146867,
+ "loss": 5.819517135620117,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.009915645234286785,
+ "learning_rate": 0.0029108483331254684,
+ "loss": 5.776939392089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.008965205401182175,
+ "learning_rate": 0.00291039288049428,
+ "loss": 5.824148178100586,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.011025927029550076,
+ "learning_rate": 0.0029099363036202515,
+ "loss": 5.832704544067383,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.010143411345779896,
+ "learning_rate": 0.0029094786028712387,
+ "loss": 5.802353382110596,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.010021256282925606,
+ "learning_rate": 0.002909019778616003,
+ "loss": 5.802136421203613,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.011356706731021404,
+ "learning_rate": 0.002908559831224211,
+ "loss": 5.807962417602539,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.010525298304855824,
+ "learning_rate": 0.002908098761066433,
+ "loss": 5.787661552429199,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.011147079057991505,
+ "learning_rate": 0.0029076365685141463,
+ "loss": 5.801015377044678,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.008777810260653496,
+ "learning_rate": 0.00290717325393973,
+ "loss": 5.779858589172363,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.011423609219491482,
+ "learning_rate": 0.002906708817716469,
+ "loss": 5.777166843414307,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.008639856241643429,
+ "learning_rate": 0.00290624326021855,
+ "loss": 5.747189521789551,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.009628898464143276,
+ "learning_rate": 0.002905776581821066,
+ "loss": 5.7666916847229,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.011988377198576927,
+ "learning_rate": 0.0029053087829000097,
+ "loss": 5.777951240539551,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.01693224348127842,
+ "learning_rate": 0.00290483986383228,
+ "loss": 5.818393707275391,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.020264796912670135,
+ "learning_rate": 0.002904369824995674,
+ "loss": 5.747611045837402,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.018283747136592865,
+ "learning_rate": 0.0029038986667688964,
+ "loss": 5.789663314819336,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.825676918029785,
+ "eval_runtime": 57.6908,
+ "eval_samples_per_second": 42.329,
+ "eval_steps_per_second": 1.335,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.011447477154433727,
+ "learning_rate": 0.002903426389531549,
+ "loss": 5.724975109100342,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.012539781630039215,
+ "learning_rate": 0.002902952993664138,
+ "loss": 5.747196674346924,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.012179837562143803,
+ "learning_rate": 0.002902478479548069,
+ "loss": 5.748995780944824,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.019896456971764565,
+ "learning_rate": 0.00290200284756565,
+ "loss": 5.751996994018555,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.022261742502450943,
+ "learning_rate": 0.002901526098100089,
+ "loss": 5.7516984939575195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02683536894619465,
+ "learning_rate": 0.0029010482315354946,
+ "loss": 5.716037750244141,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.018057072535157204,
+ "learning_rate": 0.0029005692482568747,
+ "loss": 5.778608322143555,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.011646891944110394,
+ "learning_rate": 0.002900089148650138,
+ "loss": 5.7308268547058105,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02230495773255825,
+ "learning_rate": 0.002899607933102092,
+ "loss": 5.768552780151367,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.01212464738637209,
+ "learning_rate": 0.002899125602000443,
+ "loss": 5.733015060424805,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.011764907278120518,
+ "learning_rate": 0.0028986421557337973,
+ "loss": 5.748104095458984,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.01593147963285446,
+ "learning_rate": 0.002898157594691658,
+ "loss": 5.701960563659668,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.016333773732185364,
+ "learning_rate": 0.0028976719192644277,
+ "loss": 5.716808795928955,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.017547868192195892,
+ "learning_rate": 0.002897185129843406,
+ "loss": 5.724377632141113,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.01662801019847393,
+ "learning_rate": 0.002896697226820791,
+ "loss": 5.737432956695557,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.015781691297888756,
+ "learning_rate": 0.002896208210589676,
+ "loss": 5.7376203536987305,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.014363214373588562,
+ "learning_rate": 0.0028957180815440534,
+ "loss": 5.7342610359191895,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.015048054046928883,
+ "learning_rate": 0.0028952268400788115,
+ "loss": 5.715073585510254,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.013753226026892662,
+ "learning_rate": 0.002894734486589734,
+ "loss": 5.73264217376709,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.015562563203275204,
+ "learning_rate": 0.002894241021473502,
+ "loss": 5.7072038650512695,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.01520819216966629,
+ "learning_rate": 0.0028937464451276916,
+ "loss": 5.683300018310547,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.013024724088609219,
+ "learning_rate": 0.002893250757950773,
+ "loss": 5.726273536682129,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.012760856188833714,
+ "learning_rate": 0.0028927539603421128,
+ "loss": 5.7249064445495605,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.013542579486966133,
+ "learning_rate": 0.0028922560527019723,
+ "loss": 5.70849084854126,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.014801754616200924,
+ "learning_rate": 0.002891757035431506,
+ "loss": 5.725185394287109,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.013000023551285267,
+ "learning_rate": 0.0028912569089327637,
+ "loss": 5.705799102783203,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.012919292785227299,
+ "learning_rate": 0.002890755673608688,
+ "loss": 5.674324989318848,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.015904201194643974,
+ "learning_rate": 0.002890253329863115,
+ "loss": 5.680788993835449,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.015467721968889236,
+ "learning_rate": 0.0028897498781007744,
+ "loss": 5.707531929016113,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.015157260000705719,
+ "learning_rate": 0.002889245318727288,
+ "loss": 5.689894676208496,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.014204306527972221,
+ "learning_rate": 0.00288873965214917,
+ "loss": 5.713001251220703,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.022045845165848732,
+ "learning_rate": 0.0028882328787738273,
+ "loss": 5.657631874084473,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.013380175456404686,
+ "learning_rate": 0.0028877249990095573,
+ "loss": 5.702192306518555,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.016698408871889114,
+ "learning_rate": 0.0028872160132655503,
+ "loss": 5.695206642150879,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.020001472905278206,
+ "learning_rate": 0.0028867059219518866,
+ "loss": 5.731742858886719,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.027350103482604027,
+ "learning_rate": 0.0028861947254795375,
+ "loss": 5.712314605712891,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.021862946450710297,
+ "learning_rate": 0.002885682424260365,
+ "loss": 5.697043418884277,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.019216977059841156,
+ "learning_rate": 0.0028851690187071204,
+ "loss": 5.698813438415527,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.02131095714867115,
+ "learning_rate": 0.002884654509233446,
+ "loss": 5.6918864250183105,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.016583887860178947,
+ "learning_rate": 0.0028841388962538724,
+ "loss": 5.711456298828125,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.016998684033751488,
+ "learning_rate": 0.0028836221801838194,
+ "loss": 5.675747871398926,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.01705755852162838,
+ "learning_rate": 0.0028831043614395966,
+ "loss": 5.697177410125732,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.014604352414608002,
+ "learning_rate": 0.0028825854404384,
+ "loss": 5.681950092315674,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.013561096973717213,
+ "learning_rate": 0.0028820654175983162,
+ "loss": 5.67357063293457,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.021149173378944397,
+ "learning_rate": 0.0028815442933383174,
+ "loss": 5.709477424621582,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.01431041955947876,
+ "learning_rate": 0.0028810220680782644,
+ "loss": 5.7012200355529785,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.01429709978401661,
+ "learning_rate": 0.0028804987422389045,
+ "loss": 5.689794540405273,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.032025713473558426,
+ "learning_rate": 0.0028799743162418712,
+ "loss": 5.6713762283325195,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.013168714940547943,
+ "learning_rate": 0.002879448790509686,
+ "loss": 5.685478210449219,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.014088666997849941,
+ "learning_rate": 0.0028789221654657556,
+ "loss": 5.645514965057373,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.013320053927600384,
+ "learning_rate": 0.002878394441534372,
+ "loss": 5.678678512573242,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.022180087864398956,
+ "learning_rate": 0.002877865619140712,
+ "loss": 5.660525321960449,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.01271889265626669,
+ "learning_rate": 0.002877335698710839,
+ "loss": 5.652164459228516,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.011617431417107582,
+ "learning_rate": 0.002876804680671701,
+ "loss": 5.6389946937561035,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.012040726840496063,
+ "learning_rate": 0.0028762725654511282,
+ "loss": 5.628744602203369,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.013148069381713867,
+ "learning_rate": 0.0028757393534778377,
+ "loss": 5.69610595703125,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.01826276071369648,
+ "learning_rate": 0.0028752050451814276,
+ "loss": 5.634238243103027,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.020393460988998413,
+ "learning_rate": 0.0028746696409923805,
+ "loss": 5.656968593597412,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.017072148621082306,
+ "learning_rate": 0.0028741331413420627,
+ "loss": 5.668219089508057,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.01694098301231861,
+ "learning_rate": 0.0028735955466627217,
+ "loss": 5.670073509216309,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.022760670632123947,
+ "learning_rate": 0.002873056857387488,
+ "loss": 5.649288177490234,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.03613442927598953,
+ "learning_rate": 0.0028725170739503734,
+ "loss": 5.723211288452148,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.04232329502701759,
+ "learning_rate": 0.002871976196786272,
+ "loss": 5.757920742034912,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.044210851192474365,
+ "learning_rate": 0.002871434226330959,
+ "loss": 5.702426910400391,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.044886402785778046,
+ "learning_rate": 0.0028708911630210905,
+ "loss": 5.742881774902344,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.022805195301771164,
+ "learning_rate": 0.002870347007294202,
+ "loss": 5.716353416442871,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02219395525753498,
+ "learning_rate": 0.00286980175958871,
+ "loss": 5.68503475189209,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.019240057095885277,
+ "learning_rate": 0.002869255420343911,
+ "loss": 5.700656890869141,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.01957295648753643,
+ "learning_rate": 0.0028687079899999807,
+ "loss": 5.70595645904541,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016326555982232094,
+ "learning_rate": 0.0028681594689979736,
+ "loss": 5.697229385375977,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015815885737538338,
+ "learning_rate": 0.0028676098577798234,
+ "loss": 5.7020792961120605,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.015865175053477287,
+ "learning_rate": 0.002867059156788342,
+ "loss": 5.648272514343262,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.7417449951171875,
+ "eval_runtime": 51.2263,
+ "eval_samples_per_second": 47.671,
+ "eval_steps_per_second": 1.503,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.016063183546066284,
+ "learning_rate": 0.0028665073664672185,
+ "loss": 5.65411376953125,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.014323881827294827,
+ "learning_rate": 0.0028659544872610214,
+ "loss": 5.629649639129639,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.014005238190293312,
+ "learning_rate": 0.002865400519615195,
+ "loss": 5.655135631561279,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.01385372132062912,
+ "learning_rate": 0.0028648454639760615,
+ "loss": 5.632213115692139,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.014073756523430347,
+ "learning_rate": 0.0028642893207908187,
+ "loss": 5.626431465148926,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.019551943987607956,
+ "learning_rate": 0.0028637320905075406,
+ "loss": 5.595447063446045,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.017571093514561653,
+ "learning_rate": 0.002863173773575179,
+ "loss": 5.661581039428711,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.01691303588449955,
+ "learning_rate": 0.0028626143704435585,
+ "loss": 5.641129493713379,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.01584242656826973,
+ "learning_rate": 0.0028620538815633807,
+ "loss": 5.640626907348633,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.014592235907912254,
+ "learning_rate": 0.0028614923073862213,
+ "loss": 5.611204147338867,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.016636405140161514,
+ "learning_rate": 0.0028609296483645304,
+ "loss": 5.614716053009033,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.01370833721011877,
+ "learning_rate": 0.002860365904951632,
+ "loss": 5.633011817932129,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.013218311592936516,
+ "learning_rate": 0.002859801077601725,
+ "loss": 5.588958740234375,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.016971616074442863,
+ "learning_rate": 0.0028592351667698786,
+ "loss": 5.635064125061035,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.015524905174970627,
+ "learning_rate": 0.0028586681729120393,
+ "loss": 5.611454963684082,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.014769157394766808,
+ "learning_rate": 0.0028581000964850215,
+ "loss": 5.62175178527832,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.014313343912363052,
+ "learning_rate": 0.002857530937946516,
+ "loss": 5.603628158569336,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.016748322173953056,
+ "learning_rate": 0.002856960697755083,
+ "loss": 5.617765426635742,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.013503300957381725,
+ "learning_rate": 0.0028563893763701538,
+ "loss": 5.636135578155518,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.012275655753910542,
+ "learning_rate": 0.002855816974252033,
+ "loss": 5.552533149719238,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.01288773212581873,
+ "learning_rate": 0.002855243491861894,
+ "loss": 5.620527267456055,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.013687008060514927,
+ "learning_rate": 0.002854668929661781,
+ "loss": 5.618595123291016,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.015678439289331436,
+ "learning_rate": 0.0028540932881146092,
+ "loss": 5.604031562805176,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.013567988760769367,
+ "learning_rate": 0.002853516567684162,
+ "loss": 5.549149513244629,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0157297495752573,
+ "learning_rate": 0.002852938768835092,
+ "loss": 5.618505477905273,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.011867019347846508,
+ "learning_rate": 0.0028523598920329233,
+ "loss": 5.579049110412598,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.01398563478142023,
+ "learning_rate": 0.0028517799377440443,
+ "loss": 5.597678184509277,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.015525273978710175,
+ "learning_rate": 0.0028511989064357148,
+ "loss": 5.601663589477539,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.01588466763496399,
+ "learning_rate": 0.002850616798576061,
+ "loss": 5.5650434494018555,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.014276969246566296,
+ "learning_rate": 0.0028500336146340772,
+ "loss": 5.607125282287598,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.013253689743578434,
+ "learning_rate": 0.002849449355079623,
+ "loss": 5.6299824714660645,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.01176661066710949,
+ "learning_rate": 0.0028488640203834263,
+ "loss": 5.560839653015137,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.011712565086781979,
+ "learning_rate": 0.0028482776110170813,
+ "loss": 5.618312835693359,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.013515647500753403,
+ "learning_rate": 0.002847690127453047,
+ "loss": 5.566764831542969,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.020177524536848068,
+ "learning_rate": 0.0028471015701646485,
+ "loss": 5.554571628570557,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.029197178781032562,
+ "learning_rate": 0.0028465119396260757,
+ "loss": 5.6097259521484375,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.027947140857577324,
+ "learning_rate": 0.0028459212363123824,
+ "loss": 5.65598201751709,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.021089809015393257,
+ "learning_rate": 0.0028453294606994887,
+ "loss": 5.67195987701416,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.019936271011829376,
+ "learning_rate": 0.0028447366132641773,
+ "loss": 5.609482765197754,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.0176444910466671,
+ "learning_rate": 0.002844142694484094,
+ "loss": 5.60432767868042,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.023154428228735924,
+ "learning_rate": 0.0028435477048377495,
+ "loss": 5.616270065307617,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.018391480669379234,
+ "learning_rate": 0.002842951644804515,
+ "loss": 5.600588321685791,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.013716557994484901,
+ "learning_rate": 0.002842354514864627,
+ "loss": 5.595095634460449,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.014372860081493855,
+ "learning_rate": 0.002841756315499181,
+ "loss": 5.584517478942871,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.014930845238268375,
+ "learning_rate": 0.002841157047190135,
+ "loss": 5.587005138397217,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.012631502002477646,
+ "learning_rate": 0.0028405567104203095,
+ "loss": 5.594094276428223,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.014677002094686031,
+ "learning_rate": 0.0028399553056733854,
+ "loss": 5.569582939147949,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.015558961778879166,
+ "learning_rate": 0.0028393528334339025,
+ "loss": 5.596535682678223,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.014322095550596714,
+ "learning_rate": 0.0028387492941872624,
+ "loss": 5.570324420928955,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.012913414277136326,
+ "learning_rate": 0.002838144688419726,
+ "loss": 5.573002338409424,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.01251310296356678,
+ "learning_rate": 0.0028375390166184137,
+ "loss": 5.550327777862549,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.013847017660737038,
+ "learning_rate": 0.0028369322792713026,
+ "loss": 5.54567813873291,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.013789358548820019,
+ "learning_rate": 0.0028363244768672313,
+ "loss": 5.571495056152344,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.01236631814390421,
+ "learning_rate": 0.0028357156098958954,
+ "loss": 5.569730758666992,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.013210169039666653,
+ "learning_rate": 0.002835105678847847,
+ "loss": 5.53984260559082,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.01604137383401394,
+ "learning_rate": 0.0028344946842144975,
+ "loss": 5.530598163604736,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.01708815060555935,
+ "learning_rate": 0.002833882626488114,
+ "loss": 5.529256820678711,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.017222166061401367,
+ "learning_rate": 0.0028332695061618197,
+ "loss": 5.564949035644531,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.01803218014538288,
+ "learning_rate": 0.002832655323729595,
+ "loss": 5.560049057006836,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.01658225618302822,
+ "learning_rate": 0.002832040079686276,
+ "loss": 5.553183555603027,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.016834143549203873,
+ "learning_rate": 0.002831423774527554,
+ "loss": 5.573678970336914,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.01856408454477787,
+ "learning_rate": 0.002830806408749974,
+ "loss": 5.539497375488281,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.026909375563263893,
+ "learning_rate": 0.002830187982850936,
+ "loss": 5.551433563232422,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.016046348959207535,
+ "learning_rate": 0.002829568497328696,
+ "loss": 5.566728591918945,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.018582617864012718,
+ "learning_rate": 0.0028289479526823614,
+ "loss": 5.543226718902588,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.0211537666618824,
+ "learning_rate": 0.0028283263494118943,
+ "loss": 5.560711860656738,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.013894101604819298,
+ "learning_rate": 0.0028277036880181095,
+ "loss": 5.528890132904053,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.017581885680556297,
+ "learning_rate": 0.0028270799690026725,
+ "loss": 5.57473087310791,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.01622696965932846,
+ "learning_rate": 0.002826455192868104,
+ "loss": 5.578961372375488,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.01966337114572525,
+ "learning_rate": 0.0028258293601177743,
+ "loss": 5.513942718505859,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021581366658210754,
+ "learning_rate": 0.002825202471255906,
+ "loss": 5.573669910430908,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.020468758419156075,
+ "learning_rate": 0.002824574526787571,
+ "loss": 5.516103744506836,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.626922130584717,
+ "eval_runtime": 54.2404,
+ "eval_samples_per_second": 45.022,
+ "eval_steps_per_second": 1.42,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.014597049914300442,
+ "learning_rate": 0.0028239455272186936,
+ "loss": 5.506453514099121,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.020277054980397224,
+ "learning_rate": 0.002823315473056048,
+ "loss": 5.535350799560547,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.019103925675153732,
+ "learning_rate": 0.0028226843648072562,
+ "loss": 5.522078037261963,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.07158710062503815,
+ "learning_rate": 0.0028220522029807913,
+ "loss": 5.531142234802246,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01958007924258709,
+ "learning_rate": 0.0028214189880859746,
+ "loss": 5.484055042266846,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.0185184795409441,
+ "learning_rate": 0.002820784720632976,
+ "loss": 5.511264801025391,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01976991444826126,
+ "learning_rate": 0.0028201494011328133,
+ "loss": 5.51821231842041,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.02188604325056076,
+ "learning_rate": 0.002819513030097352,
+ "loss": 5.483095169067383,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.019405223429203033,
+ "learning_rate": 0.002818875608039305,
+ "loss": 5.539701461791992,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.016907615587115288,
+ "learning_rate": 0.0028182371354722306,
+ "loss": 5.518251419067383,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.02160538360476494,
+ "learning_rate": 0.0028175976129105364,
+ "loss": 5.509786605834961,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.023125721141695976,
+ "learning_rate": 0.002816957040869473,
+ "loss": 5.535561561584473,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.022233380004763603,
+ "learning_rate": 0.0028163154198651372,
+ "loss": 5.553831100463867,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.017227552831172943,
+ "learning_rate": 0.002815672750414473,
+ "loss": 5.547080993652344,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.017122473567724228,
+ "learning_rate": 0.0028150290330352665,
+ "loss": 5.493778228759766,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.020641913637518883,
+ "learning_rate": 0.0028143842682461498,
+ "loss": 5.536645889282227,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.017383966594934464,
+ "learning_rate": 0.0028137384565665967,
+ "loss": 5.506233215332031,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.017969606444239616,
+ "learning_rate": 0.002813091598516928,
+ "loss": 5.524741172790527,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.01936613954603672,
+ "learning_rate": 0.0028124436946183047,
+ "loss": 5.460989952087402,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.017492495477199554,
+ "learning_rate": 0.0028117947453927313,
+ "loss": 5.487619876861572,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.018786396831274033,
+ "learning_rate": 0.0028111447513630545,
+ "loss": 5.516829013824463,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.020149117335677147,
+ "learning_rate": 0.0028104937130529624,
+ "loss": 5.460541725158691,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.024668103083968163,
+ "learning_rate": 0.0028098416309869857,
+ "loss": 5.497687339782715,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.015888608992099762,
+ "learning_rate": 0.002809188505690494,
+ "loss": 5.494459629058838,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.019327623769640923,
+ "learning_rate": 0.0028085343376896997,
+ "loss": 5.486351490020752,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.026855193078517914,
+ "learning_rate": 0.0028078791275116532,
+ "loss": 5.509825706481934,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.026340370997786522,
+ "learning_rate": 0.002807222875684246,
+ "loss": 5.527161598205566,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.02200835384428501,
+ "learning_rate": 0.002806565582736208,
+ "loss": 5.507146835327148,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.020420534536242485,
+ "learning_rate": 0.0028059072491971085,
+ "loss": 5.5162506103515625,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02493203990161419,
+ "learning_rate": 0.0028052478755973545,
+ "loss": 5.5252556800842285,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.021293973550200462,
+ "learning_rate": 0.0028045874624681925,
+ "loss": 5.523589134216309,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.016593236476182938,
+ "learning_rate": 0.0028039260103417045,
+ "loss": 5.500586032867432,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.018857723101973534,
+ "learning_rate": 0.002803263519750811,
+ "loss": 5.52933406829834,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.015930436551570892,
+ "learning_rate": 0.0028025999912292677,
+ "loss": 5.48813533782959,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.019427936524152756,
+ "learning_rate": 0.002801935425311669,
+ "loss": 5.46573543548584,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.018123053014278412,
+ "learning_rate": 0.0028012698225334427,
+ "loss": 5.499445915222168,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.02035410702228546,
+ "learning_rate": 0.002800603183430853,
+ "loss": 5.513540267944336,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.025838350877165794,
+ "learning_rate": 0.0027999355085409996,
+ "loss": 5.565637588500977,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.029621612280607224,
+ "learning_rate": 0.002799266798401816,
+ "loss": 5.5619215965271,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.028948379680514336,
+ "learning_rate": 0.0027985970535520685,
+ "loss": 5.551331520080566,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.022135930135846138,
+ "learning_rate": 0.0027979262745313604,
+ "loss": 5.527722358703613,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.028884069994091988,
+ "learning_rate": 0.0027972544618801256,
+ "loss": 5.554599761962891,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.02499174140393734,
+ "learning_rate": 0.00279658161613963,
+ "loss": 5.543351173400879,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.021131042391061783,
+ "learning_rate": 0.0027959077378519755,
+ "loss": 5.496039390563965,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.02957489714026451,
+ "learning_rate": 0.002795232827560092,
+ "loss": 5.552300453186035,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.022832119837403297,
+ "learning_rate": 0.0027945568858077433,
+ "loss": 5.488651275634766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.02074701525270939,
+ "learning_rate": 0.002793879913139523,
+ "loss": 5.491244792938232,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.019919803366065025,
+ "learning_rate": 0.002793201910100856,
+ "loss": 5.483855247497559,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.01858169585466385,
+ "learning_rate": 0.002792522877237998,
+ "loss": 5.50787353515625,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02085857465863228,
+ "learning_rate": 0.002791842815098032,
+ "loss": 5.512872695922852,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.01407450158149004,
+ "learning_rate": 0.002791161724228872,
+ "loss": 5.472590446472168,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.01672944612801075,
+ "learning_rate": 0.002790479605179262,
+ "loss": 5.469183921813965,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.016565700992941856,
+ "learning_rate": 0.0027897964584987713,
+ "loss": 5.493677139282227,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.017017360776662827,
+ "learning_rate": 0.0027891122847377996,
+ "loss": 5.464931488037109,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.019842486828565598,
+ "learning_rate": 0.002788427084447574,
+ "loss": 5.461293697357178,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.016210904344916344,
+ "learning_rate": 0.0027877408581801467,
+ "loss": 5.454716682434082,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.02064039371907711,
+ "learning_rate": 0.002787053606488398,
+ "loss": 5.466576099395752,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.018198711797595024,
+ "learning_rate": 0.0027863653299260356,
+ "loss": 5.447747230529785,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.015937626361846924,
+ "learning_rate": 0.00278567602904759,
+ "loss": 5.5055036544799805,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.01687259040772915,
+ "learning_rate": 0.002784985704408419,
+ "loss": 5.422250747680664,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.02622046135365963,
+ "learning_rate": 0.002784294356564705,
+ "loss": 5.466894626617432,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.018672186881303787,
+ "learning_rate": 0.0027836019860734535,
+ "loss": 5.48300838470459,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.02166241593658924,
+ "learning_rate": 0.0027829085934924958,
+ "loss": 5.477658271789551,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018426023423671722,
+ "learning_rate": 0.0027822141793804856,
+ "loss": 5.487186431884766,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.022229891270399094,
+ "learning_rate": 0.0027815187442968995,
+ "loss": 5.488579273223877,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.026930972933769226,
+ "learning_rate": 0.002780822288802037,
+ "loss": 5.492000102996826,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.02733265422284603,
+ "learning_rate": 0.00278012481345702,
+ "loss": 5.4663801193237305,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.02993253618478775,
+ "learning_rate": 0.0027794263188237922,
+ "loss": 5.543368816375732,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03152153640985489,
+ "learning_rate": 0.002778726805465118,
+ "loss": 5.565644264221191,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.031395044177770615,
+ "learning_rate": 0.002778026273944581,
+ "loss": 5.553359508514404,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.026603225618600845,
+ "learning_rate": 0.002777324724826589,
+ "loss": 5.552347183227539,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.024152139201760292,
+ "learning_rate": 0.002776622158676366,
+ "loss": 5.493257522583008,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.602636814117432,
+ "eval_runtime": 54.6199,
+ "eval_samples_per_second": 44.709,
+ "eval_steps_per_second": 1.41,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.02089761756360531,
+ "learning_rate": 0.0027759185760599576,
+ "loss": 5.461747646331787,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.018181778490543365,
+ "learning_rate": 0.0027752139775442278,
+ "loss": 5.492741584777832,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.020166238769888878,
+ "learning_rate": 0.0027745083636968575,
+ "loss": 5.476003646850586,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.015491694211959839,
+ "learning_rate": 0.0027738017350863488,
+ "loss": 5.425764560699463,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.022291697561740875,
+ "learning_rate": 0.0027730940922820182,
+ "loss": 5.45866584777832,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.018074683845043182,
+ "learning_rate": 0.002772385435854002,
+ "loss": 5.465143203735352,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.016872268170118332,
+ "learning_rate": 0.002771675766373251,
+ "loss": 5.446003437042236,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.023103339597582817,
+ "learning_rate": 0.0027709650844115336,
+ "loss": 5.411036491394043,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.027957158163189888,
+ "learning_rate": 0.0027702533905414324,
+ "loss": 5.452442169189453,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.019935399293899536,
+ "learning_rate": 0.002769540685336348,
+ "loss": 5.412145614624023,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.014351371675729752,
+ "learning_rate": 0.0027688269693704926,
+ "loss": 5.434524059295654,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.0152254244312644,
+ "learning_rate": 0.0027681122432188948,
+ "loss": 5.428544998168945,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.015071330592036247,
+ "learning_rate": 0.0027673965074573965,
+ "loss": 5.459929466247559,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.023415958508849144,
+ "learning_rate": 0.0027666797626626537,
+ "loss": 5.447159767150879,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.017228802666068077,
+ "learning_rate": 0.0027659620094121337,
+ "loss": 5.442273139953613,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.014184504747390747,
+ "learning_rate": 0.002765243248284118,
+ "loss": 5.448305130004883,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.024384526535868645,
+ "learning_rate": 0.0027645234798576988,
+ "loss": 5.4076337814331055,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.043152231723070145,
+ "learning_rate": 0.0027638027047127813,
+ "loss": 5.45344352722168,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.02889758162200451,
+ "learning_rate": 0.002763080923430081,
+ "loss": 5.489194393157959,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.02686968818306923,
+ "learning_rate": 0.002762358136591123,
+ "loss": 5.493659496307373,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.020467868074774742,
+ "learning_rate": 0.002761634344778245,
+ "loss": 5.410489559173584,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.019759224727749825,
+ "learning_rate": 0.002760909548574592,
+ "loss": 5.439385890960693,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.0233843382447958,
+ "learning_rate": 0.0027601837485641194,
+ "loss": 5.445036888122559,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.031753458082675934,
+ "learning_rate": 0.0027594569453315916,
+ "loss": 5.503895282745361,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.02089349366724491,
+ "learning_rate": 0.002758729139462581,
+ "loss": 5.45689582824707,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.03339357674121857,
+ "learning_rate": 0.002758000331543466,
+ "loss": 5.522361755371094,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019328996539115906,
+ "learning_rate": 0.002757270522161436,
+ "loss": 5.441373825073242,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019863856956362724,
+ "learning_rate": 0.002756539711904484,
+ "loss": 5.4633283615112305,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.02141754701733589,
+ "learning_rate": 0.0027558079013614116,
+ "loss": 5.453215599060059,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016591915860772133,
+ "learning_rate": 0.002755075091121824,
+ "loss": 5.463696479797363,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.01955028623342514,
+ "learning_rate": 0.0027543412817761345,
+ "loss": 5.411355018615723,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.016791662201285362,
+ "learning_rate": 0.0027536064739155596,
+ "loss": 5.459614276885986,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.019661981612443924,
+ "learning_rate": 0.00275287066813212,
+ "loss": 5.440699577331543,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.01927965320646763,
+ "learning_rate": 0.002752133865018642,
+ "loss": 5.409865856170654,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.016483580693602562,
+ "learning_rate": 0.002751396065168754,
+ "loss": 5.390999794006348,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016264773905277252,
+ "learning_rate": 0.0027506572691768888,
+ "loss": 5.418543815612793,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.01958460919559002,
+ "learning_rate": 0.00274991747763828,
+ "loss": 5.403289794921875,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.01861695572733879,
+ "learning_rate": 0.002749176691148965,
+ "loss": 5.431911468505859,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.029594263061881065,
+ "learning_rate": 0.002748434910305782,
+ "loss": 5.399242401123047,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.028574613854289055,
+ "learning_rate": 0.0027476921357063697,
+ "loss": 5.440446853637695,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.028612585738301277,
+ "learning_rate": 0.0027469483679491693,
+ "loss": 5.410346031188965,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.020704898983240128,
+ "learning_rate": 0.00274620360763342,
+ "loss": 5.44694709777832,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.020542364567518234,
+ "learning_rate": 0.002745457855359163,
+ "loss": 5.402653694152832,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.015762662515044212,
+ "learning_rate": 0.0027447111117272363,
+ "loss": 5.396594047546387,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022096984088420868,
+ "learning_rate": 0.002743963377339278,
+ "loss": 5.398744106292725,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.018791764974594116,
+ "learning_rate": 0.002743214652797724,
+ "loss": 5.397758483886719,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01642526686191559,
+ "learning_rate": 0.0027424649387058093,
+ "loss": 5.411459922790527,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.017931703478097916,
+ "learning_rate": 0.0027417142356675637,
+ "loss": 5.390264511108398,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018819602206349373,
+ "learning_rate": 0.0027409625442878153,
+ "loss": 5.407539367675781,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.019389411434531212,
+ "learning_rate": 0.0027402098651721886,
+ "loss": 5.386474609375,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.013645983301103115,
+ "learning_rate": 0.002739456198927104,
+ "loss": 5.373845100402832,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.013846895657479763,
+ "learning_rate": 0.002738701546159776,
+ "loss": 5.387155055999756,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013553125783801079,
+ "learning_rate": 0.002737945907478215,
+ "loss": 5.377801895141602,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.012771612964570522,
+ "learning_rate": 0.002737189283491225,
+ "loss": 5.344066143035889,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.020465301349759102,
+ "learning_rate": 0.002736431674808405,
+ "loss": 5.382771968841553,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.014798106625676155,
+ "learning_rate": 0.0027356730820401465,
+ "loss": 5.370522499084473,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.01566634140908718,
+ "learning_rate": 0.0027349135057976332,
+ "loss": 5.351980209350586,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.017485102638602257,
+ "learning_rate": 0.002734152946692843,
+ "loss": 5.355443477630615,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.016646096482872963,
+ "learning_rate": 0.0027333914053385444,
+ "loss": 5.34829044342041,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.017293352633714676,
+ "learning_rate": 0.0027326288823482964,
+ "loss": 5.3519487380981445,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.018873002380132675,
+ "learning_rate": 0.002731865378336452,
+ "loss": 5.348442077636719,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.015504784882068634,
+ "learning_rate": 0.002731100893918151,
+ "loss": 5.341490268707275,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.014270870015025139,
+ "learning_rate": 0.0027303354297093254,
+ "loss": 5.359251022338867,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.018820680677890778,
+ "learning_rate": 0.002729568986326695,
+ "loss": 5.390280723571777,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.021008187904953957,
+ "learning_rate": 0.0027288015643877708,
+ "loss": 5.338380336761475,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.016266655176877975,
+ "learning_rate": 0.0027280331645108498,
+ "loss": 5.347043037414551,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.01831045374274254,
+ "learning_rate": 0.0027272637873150185,
+ "loss": 5.326537132263184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02073794975876808,
+ "learning_rate": 0.002726493433420149,
+ "loss": 5.3607025146484375,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.022534754127264023,
+ "learning_rate": 0.0027257221034469033,
+ "loss": 5.353292465209961,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.021627362817525864,
+ "learning_rate": 0.002724949798016727,
+ "loss": 5.373323917388916,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.018226243555545807,
+ "learning_rate": 0.0027241765177518523,
+ "loss": 5.344578266143799,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.02032875455915928,
+ "learning_rate": 0.002723402263275298,
+ "loss": 5.329159259796143,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.44969367980957,
+ "eval_runtime": 51.2294,
+ "eval_samples_per_second": 47.668,
+ "eval_steps_per_second": 1.503,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.01613701693713665,
+ "learning_rate": 0.0027226270352108664,
+ "loss": 5.3024678230285645,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01940973661839962,
+ "learning_rate": 0.0027218508341831445,
+ "loss": 5.270181179046631,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.015840493142604828,
+ "learning_rate": 0.002721073660817504,
+ "loss": 5.261706352233887,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.017390675842761993,
+ "learning_rate": 0.0027202955157401,
+ "loss": 5.262207984924316,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.015184012241661549,
+ "learning_rate": 0.0027195163995778686,
+ "loss": 5.297155380249023,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.013417073525488377,
+ "learning_rate": 0.00271873631295853,
+ "loss": 5.279838562011719,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.013024317100644112,
+ "learning_rate": 0.0027179552565105864,
+ "loss": 5.265446186065674,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.01426338218152523,
+ "learning_rate": 0.002717173230863321,
+ "loss": 5.284180164337158,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.019604040309786797,
+ "learning_rate": 0.002716390236646797,
+ "loss": 5.286281108856201,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.019039476290345192,
+ "learning_rate": 0.0027156062744918587,
+ "loss": 5.290433406829834,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.019284334033727646,
+ "learning_rate": 0.0027148213450301313,
+ "loss": 5.262433052062988,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.0145154669880867,
+ "learning_rate": 0.0027140354488940175,
+ "loss": 5.24387264251709,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.014776417054235935,
+ "learning_rate": 0.0027132485867166992,
+ "loss": 5.282435417175293,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.012711729854345322,
+ "learning_rate": 0.0027124607591321377,
+ "loss": 5.246735572814941,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.013394835405051708,
+ "learning_rate": 0.0027116719667750716,
+ "loss": 5.270739555358887,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.013468850404024124,
+ "learning_rate": 0.002710882210281016,
+ "loss": 5.27161169052124,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.030262840911746025,
+ "learning_rate": 0.002710091490286264,
+ "loss": 5.256673812866211,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.013560288585722446,
+ "learning_rate": 0.0027092998074278846,
+ "loss": 5.253674507141113,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014592813327908516,
+ "learning_rate": 0.002708507162343721,
+ "loss": 5.25804328918457,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014328169636428356,
+ "learning_rate": 0.0027077135556723953,
+ "loss": 5.235117435455322,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.01277366653084755,
+ "learning_rate": 0.0027069189880533,
+ "loss": 5.2053985595703125,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.012948785908520222,
+ "learning_rate": 0.0027061234601266043,
+ "loss": 5.214653015136719,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.014892828650772572,
+ "learning_rate": 0.002705326972533252,
+ "loss": 5.227547645568848,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01427230704575777,
+ "learning_rate": 0.002704529525914957,
+ "loss": 5.241065502166748,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.01650116965174675,
+ "learning_rate": 0.0027037311209142092,
+ "loss": 5.227713108062744,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.019728578627109528,
+ "learning_rate": 0.0027029317581742677,
+ "loss": 5.2624921798706055,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.023847149685025215,
+ "learning_rate": 0.0027021314383391657,
+ "loss": 5.278811454772949,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.020690469071269035,
+ "learning_rate": 0.0027013301620537056,
+ "loss": 5.238919258117676,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.021211959421634674,
+ "learning_rate": 0.0027005279299634616,
+ "loss": 5.253684997558594,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.025854922831058502,
+ "learning_rate": 0.002699724742714778,
+ "loss": 5.2538652420043945,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.0215485580265522,
+ "learning_rate": 0.0026989206009547678,
+ "loss": 5.224540710449219,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.019827792420983315,
+ "learning_rate": 0.0026981155053313137,
+ "loss": 5.220674514770508,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.017689093947410583,
+ "learning_rate": 0.002697309456493066,
+ "loss": 5.228511810302734,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.022564806044101715,
+ "learning_rate": 0.002696502455089445,
+ "loss": 5.295933723449707,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.020070280879735947,
+ "learning_rate": 0.0026956945017706357,
+ "loss": 5.230920791625977,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.025744661688804626,
+ "learning_rate": 0.002694885597187593,
+ "loss": 5.296453952789307,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.021263007074594498,
+ "learning_rate": 0.0026940757419920357,
+ "loss": 5.220481872558594,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.017577245831489563,
+ "learning_rate": 0.0026932649368364495,
+ "loss": 5.231488227844238,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.015562832355499268,
+ "learning_rate": 0.002692453182374086,
+ "loss": 5.175224304199219,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.014115403406322002,
+ "learning_rate": 0.0026916404792589606,
+ "loss": 5.228387832641602,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.012863830663263798,
+ "learning_rate": 0.002690826828145854,
+ "loss": 5.186727523803711,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.014151126146316528,
+ "learning_rate": 0.002690012229690309,
+ "loss": 5.1694817543029785,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.01316932775080204,
+ "learning_rate": 0.0026891966845486353,
+ "loss": 5.19931697845459,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.011340726166963577,
+ "learning_rate": 0.0026883801933779003,
+ "loss": 5.238619804382324,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.012616089545190334,
+ "learning_rate": 0.0026875627568359378,
+ "loss": 5.183376789093018,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.012274038046598434,
+ "learning_rate": 0.002686744375581341,
+ "loss": 5.199946403503418,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.014990398660302162,
+ "learning_rate": 0.002685925050273465,
+ "loss": 5.168451309204102,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.011399757117033005,
+ "learning_rate": 0.0026851047815724255,
+ "loss": 5.173828125,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.013638390228152275,
+ "learning_rate": 0.0026842835701390982,
+ "loss": 5.142702102661133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.014472718350589275,
+ "learning_rate": 0.0026834614166351177,
+ "loss": 5.173436164855957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.014509808272123337,
+ "learning_rate": 0.002682638321722879,
+ "loss": 5.173975467681885,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.013576583005487919,
+ "learning_rate": 0.0026818142860655352,
+ "loss": 5.1435089111328125,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.012914472259581089,
+ "learning_rate": 0.0026809893103269954,
+ "loss": 5.151381492614746,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.013803352601826191,
+ "learning_rate": 0.0026801633951719288,
+ "loss": 5.133159637451172,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.018488213419914246,
+ "learning_rate": 0.0026793365412657606,
+ "loss": 5.183903694152832,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.01843969337642193,
+ "learning_rate": 0.002678508749274671,
+ "loss": 5.1660356521606445,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.017181839793920517,
+ "learning_rate": 0.0026776800198655983,
+ "loss": 5.179847717285156,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0148160420358181,
+ "learning_rate": 0.0026768503537062333,
+ "loss": 5.149198532104492,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.013251296244561672,
+ "learning_rate": 0.0026760197514650247,
+ "loss": 5.192000389099121,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.010746927931904793,
+ "learning_rate": 0.0026751882138111717,
+ "loss": 5.138680458068848,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.012316994369029999,
+ "learning_rate": 0.002674355741414631,
+ "loss": 5.169965744018555,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.01191797386854887,
+ "learning_rate": 0.0026735223349461106,
+ "loss": 5.128267288208008,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.013525084592401981,
+ "learning_rate": 0.0026726879950770698,
+ "loss": 5.124795436859131,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.014843370765447617,
+ "learning_rate": 0.002671852722479721,
+ "loss": 5.1497087478637695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.015285412780940533,
+ "learning_rate": 0.00267101651782703,
+ "loss": 5.093118190765381,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.013555832207202911,
+ "learning_rate": 0.0026701793817927107,
+ "loss": 5.13249397277832,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.010569742880761623,
+ "learning_rate": 0.0026693413150512273,
+ "loss": 5.084077835083008,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.011483813636004925,
+ "learning_rate": 0.0026685023182777975,
+ "loss": 5.128049850463867,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.013724366202950478,
+ "learning_rate": 0.0026676623921483844,
+ "loss": 5.157539367675781,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.014337432570755482,
+ "learning_rate": 0.0026668215373397016,
+ "loss": 5.149948596954346,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.014101454056799412,
+ "learning_rate": 0.00266597975452921,
+ "loss": 5.117250442504883,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.016004368662834167,
+ "learning_rate": 0.0026651370443951197,
+ "loss": 5.143221855163574,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.249833583831787,
+ "eval_runtime": 51.5543,
+ "eval_samples_per_second": 47.368,
+ "eval_steps_per_second": 1.494,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.01297403872013092,
+ "learning_rate": 0.0026642934076163867,
+ "loss": 5.089287281036377,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.012867718935012817,
+ "learning_rate": 0.0026634488448727144,
+ "loss": 5.0784101486206055,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.014600927010178566,
+ "learning_rate": 0.0026626033568445504,
+ "loss": 5.050220489501953,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.016619572415947914,
+ "learning_rate": 0.0026617569442130905,
+ "loss": 5.079303741455078,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.015560587868094444,
+ "learning_rate": 0.0026609096076602736,
+ "loss": 5.072385311126709,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.01774057187139988,
+ "learning_rate": 0.0026600613478687833,
+ "loss": 5.122529983520508,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.019971400499343872,
+ "learning_rate": 0.0026592121655220472,
+ "loss": 5.085966110229492,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.017667256295681,
+ "learning_rate": 0.0026583620613042365,
+ "loss": 5.078432559967041,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.022369660437107086,
+ "learning_rate": 0.002657511035900264,
+ "loss": 5.0630903244018555,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.017964458093047142,
+ "learning_rate": 0.002656659089995787,
+ "loss": 5.0846076011657715,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.021769145503640175,
+ "learning_rate": 0.002655806224277201,
+ "loss": 5.081344127655029,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.02487620897591114,
+ "learning_rate": 0.0026549524394316464,
+ "loss": 5.106287956237793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.022516218945384026,
+ "learning_rate": 0.0026540977361470012,
+ "loss": 5.128098487854004,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.02099926583468914,
+ "learning_rate": 0.0026532421151118845,
+ "loss": 5.103562355041504,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.01899387687444687,
+ "learning_rate": 0.002652385577015655,
+ "loss": 5.062675476074219,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.016801996156573296,
+ "learning_rate": 0.002651528122548409,
+ "loss": 5.096002578735352,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.014009200967848301,
+ "learning_rate": 0.0026506697524009836,
+ "loss": 5.075201034545898,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.014653858728706837,
+ "learning_rate": 0.002649810467264951,
+ "loss": 5.090478897094727,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.014918150380253792,
+ "learning_rate": 0.0026489502678326226,
+ "loss": 5.070797920227051,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.01327288057655096,
+ "learning_rate": 0.0026480891547970443,
+ "loss": 5.030500411987305,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.012885140255093575,
+ "learning_rate": 0.002647227128852,
+ "loss": 5.0845537185668945,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.012848512269556522,
+ "learning_rate": 0.002646364190692009,
+ "loss": 5.079484939575195,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.01149789709597826,
+ "learning_rate": 0.002645500341012324,
+ "loss": 5.036316394805908,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.010360757820308208,
+ "learning_rate": 0.0026446355805089337,
+ "loss": 5.04129695892334,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.010426290333271027,
+ "learning_rate": 0.00264376990987856,
+ "loss": 5.051270484924316,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.01018153689801693,
+ "learning_rate": 0.002642903329818658,
+ "loss": 5.102342128753662,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.011879283003509045,
+ "learning_rate": 0.0026420358410274153,
+ "loss": 5.053501129150391,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.010418329387903214,
+ "learning_rate": 0.002641167444203752,
+ "loss": 5.030653476715088,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.012302013114094734,
+ "learning_rate": 0.0026402981400473202,
+ "loss": 5.0402069091796875,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.011392916552722454,
+ "learning_rate": 0.002639427929258502,
+ "loss": 5.024111747741699,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.013075058348476887,
+ "learning_rate": 0.0026385568125384114,
+ "loss": 5.021731376647949,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015349832363426685,
+ "learning_rate": 0.0026376847905888903,
+ "loss": 5.03192138671875,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01588309183716774,
+ "learning_rate": 0.0026368118641125113,
+ "loss": 4.966363906860352,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.014378590509295464,
+ "learning_rate": 0.0026359380338125764,
+ "loss": 5.070773124694824,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.012306655757129192,
+ "learning_rate": 0.002635063300393114,
+ "loss": 5.025704860687256,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01212004292756319,
+ "learning_rate": 0.002634187664558881,
+ "loss": 4.965104103088379,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013538849540054798,
+ "learning_rate": 0.0026333111270153624,
+ "loss": 5.029810905456543,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.012052906677126884,
+ "learning_rate": 0.0026324336884687672,
+ "loss": 5.0262908935546875,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.011368660256266594,
+ "learning_rate": 0.002631555349626033,
+ "loss": 5.001485824584961,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.010215395130217075,
+ "learning_rate": 0.0026306761111948205,
+ "loss": 5.019739151000977,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.011449878104031086,
+ "learning_rate": 0.0026297959738835167,
+ "loss": 4.9816083908081055,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.012381811626255512,
+ "learning_rate": 0.0026289149384012326,
+ "loss": 5.025032043457031,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.00974641740322113,
+ "learning_rate": 0.002628033005457802,
+ "loss": 5.000177383422852,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.009515874087810516,
+ "learning_rate": 0.0026271501757637833,
+ "loss": 4.996979713439941,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.011125455610454082,
+ "learning_rate": 0.0026262664500304555,
+ "loss": 4.996796607971191,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.013290468603372574,
+ "learning_rate": 0.0026253818289698207,
+ "loss": 5.053703308105469,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.014650283381342888,
+ "learning_rate": 0.002624496313294602,
+ "loss": 5.001321792602539,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014537827111780643,
+ "learning_rate": 0.002623609903718244,
+ "loss": 5.024584770202637,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.012908566743135452,
+ "learning_rate": 0.0026227226009549105,
+ "loss": 5.042878150939941,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.01446808222681284,
+ "learning_rate": 0.002621834405719485,
+ "loss": 5.020857810974121,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.01535153016448021,
+ "learning_rate": 0.0026209453187275708,
+ "loss": 5.0056304931640625,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01815616339445114,
+ "learning_rate": 0.0026200553406954892,
+ "loss": 5.0467376708984375,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.023135516792535782,
+ "learning_rate": 0.0026191644723402795,
+ "loss": 5.033995628356934,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.019682930782437325,
+ "learning_rate": 0.002618272714379698,
+ "loss": 5.008149147033691,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.018521904945373535,
+ "learning_rate": 0.002617380067532217,
+ "loss": 5.007076263427734,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.02118663489818573,
+ "learning_rate": 0.0026164865325170278,
+ "loss": 5.05519962310791,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024684788659214973,
+ "learning_rate": 0.0026155921100540344,
+ "loss": 5.027571201324463,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.02609819918870926,
+ "learning_rate": 0.0026146968008638567,
+ "loss": 5.10361385345459,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.025958053767681122,
+ "learning_rate": 0.00261380060566783,
+ "loss": 5.0911455154418945,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.017067518085241318,
+ "learning_rate": 0.002612903525188002,
+ "loss": 5.045783042907715,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.016510486602783203,
+ "learning_rate": 0.0026120055601471345,
+ "loss": 5.069464683532715,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015552333556115627,
+ "learning_rate": 0.0026111067112687012,
+ "loss": 5.024248123168945,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.013432685285806656,
+ "learning_rate": 0.0026102069792768886,
+ "loss": 5.004837989807129,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.012300201691687107,
+ "learning_rate": 0.0026093063648965943,
+ "loss": 5.031093597412109,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.01197744719684124,
+ "learning_rate": 0.002608404868853428,
+ "loss": 4.965237617492676,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.01100246049463749,
+ "learning_rate": 0.002607502491873708,
+ "loss": 5.04719352722168,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.010111447423696518,
+ "learning_rate": 0.0026065992346844626,
+ "loss": 4.986410617828369,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.010669790208339691,
+ "learning_rate": 0.002605695098013431,
+ "loss": 4.9825663566589355,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.010805477388203144,
+ "learning_rate": 0.0026047900825890593,
+ "loss": 4.991506576538086,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.010886074975132942,
+ "learning_rate": 0.002603884189140502,
+ "loss": 4.987531661987305,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.010735839605331421,
+ "learning_rate": 0.0026029774183976206,
+ "loss": 4.982756614685059,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.011092593893408775,
+ "learning_rate": 0.002602069771090985,
+ "loss": 4.994824409484863,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 5.13364315032959,
+ "eval_runtime": 50.4195,
+ "eval_samples_per_second": 48.434,
+ "eval_steps_per_second": 1.527,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.01005905494093895,
+ "learning_rate": 0.0026011612479518694,
+ "loss": 4.918412685394287,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.009904682636260986,
+ "learning_rate": 0.002600251849712255,
+ "loss": 4.924630165100098,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.009543103165924549,
+ "learning_rate": 0.0025993415771048273,
+ "loss": 4.940309524536133,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.011031470261514187,
+ "learning_rate": 0.0025984304308629765,
+ "loss": 4.924609184265137,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.009511022828519344,
+ "learning_rate": 0.002597518411720796,
+ "loss": 4.9102325439453125,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.00886621605604887,
+ "learning_rate": 0.0025966055204130847,
+ "loss": 4.9236907958984375,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.010057039558887482,
+ "learning_rate": 0.002595691757675341,
+ "loss": 4.908633232116699,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.01098555512726307,
+ "learning_rate": 0.0025947771242437684,
+ "loss": 4.935927391052246,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.008800056762993336,
+ "learning_rate": 0.0025938616208552694,
+ "loss": 4.873291969299316,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.009747766889631748,
+ "learning_rate": 0.0025929452482474497,
+ "loss": 4.954436302185059,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.010785684920847416,
+ "learning_rate": 0.002592028007158613,
+ "loss": 4.924927711486816,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.012620717287063599,
+ "learning_rate": 0.0025911098983277645,
+ "loss": 4.953606605529785,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.014428738504648209,
+ "learning_rate": 0.002590190922494607,
+ "loss": 4.951706886291504,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.014496048912405968,
+ "learning_rate": 0.0025892710803995444,
+ "loss": 4.941249847412109,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.0109238987788558,
+ "learning_rate": 0.0025883503727836764,
+ "loss": 4.939975261688232,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.012271072715520859,
+ "learning_rate": 0.0025874288003887996,
+ "loss": 4.880942344665527,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.016302887350320816,
+ "learning_rate": 0.0025865063639574086,
+ "loss": 4.931602478027344,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.016137251630425453,
+ "learning_rate": 0.0025855830642326945,
+ "loss": 4.918013095855713,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.014732961542904377,
+ "learning_rate": 0.002584658901958542,
+ "loss": 4.927577495574951,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.016387946903705597,
+ "learning_rate": 0.0025837338778795325,
+ "loss": 4.930819034576416,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.017146151512861252,
+ "learning_rate": 0.002582807992740941,
+ "loss": 4.928377151489258,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.016153613105416298,
+ "learning_rate": 0.0025818812472887375,
+ "loss": 4.975607395172119,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.013386047445237637,
+ "learning_rate": 0.0025809536422695825,
+ "loss": 4.9434709548950195,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.012328253127634525,
+ "learning_rate": 0.0025800251784308314,
+ "loss": 4.917585372924805,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.012000586837530136,
+ "learning_rate": 0.0025790958565205316,
+ "loss": 4.9137725830078125,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.009463997557759285,
+ "learning_rate": 0.002578165677287419,
+ "loss": 4.910362243652344,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.010591360740363598,
+ "learning_rate": 0.0025772346414809242,
+ "loss": 4.916032314300537,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.010429762303829193,
+ "learning_rate": 0.002576302749851164,
+ "loss": 4.934561729431152,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.12526583671569824,
+ "learning_rate": 0.0025753700031489486,
+ "loss": 4.897515773773193,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01368892751634121,
+ "learning_rate": 0.0025744364021257743,
+ "loss": 4.94814920425415,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.016603967174887657,
+ "learning_rate": 0.0025735019475338254,
+ "loss": 4.8996758460998535,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.06110106781125069,
+ "learning_rate": 0.0025725666401259775,
+ "loss": 4.946794509887695,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.018292423337697983,
+ "learning_rate": 0.002571630480655789,
+ "loss": 4.958992958068848,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.01852245442569256,
+ "learning_rate": 0.0025706934698775073,
+ "loss": 4.942556381225586,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.0195649154484272,
+ "learning_rate": 0.0025697556085460646,
+ "loss": 4.895773887634277,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.017820244655013084,
+ "learning_rate": 0.00256881689741708,
+ "loss": 4.915250778198242,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.016814742237329483,
+ "learning_rate": 0.002567877337246855,
+ "loss": 4.9521260261535645,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.015319781377911568,
+ "learning_rate": 0.002566936928792377,
+ "loss": 4.899402141571045,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.01607401855289936,
+ "learning_rate": 0.0025659956728113146,
+ "loss": 4.909913063049316,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.015564955770969391,
+ "learning_rate": 0.002565053570062023,
+ "loss": 4.947719573974609,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.017502037808299065,
+ "learning_rate": 0.002564110621303536,
+ "loss": 4.95107364654541,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.014791073277592659,
+ "learning_rate": 0.002563166827295571,
+ "loss": 4.8995866775512695,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.012700124643743038,
+ "learning_rate": 0.002562222188798525,
+ "loss": 4.922100067138672,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.011235272511839867,
+ "learning_rate": 0.0025612767065734763,
+ "loss": 4.884456634521484,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.013009768910706043,
+ "learning_rate": 0.002560330381382183,
+ "loss": 4.869503021240234,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.0102638378739357,
+ "learning_rate": 0.0025593832139870834,
+ "loss": 4.934551239013672,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.011081870645284653,
+ "learning_rate": 0.0025584352051512913,
+ "loss": 4.937268257141113,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.010681218467652798,
+ "learning_rate": 0.002557486355638601,
+ "loss": 4.920559883117676,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.010962890461087227,
+ "learning_rate": 0.002556536666213484,
+ "loss": 4.936082363128662,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.010790379717946053,
+ "learning_rate": 0.0025555861376410868,
+ "loss": 4.9506449699401855,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.010659157298505306,
+ "learning_rate": 0.0025546347706872346,
+ "loss": 4.872883319854736,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.012962655164301395,
+ "learning_rate": 0.0025536825661184254,
+ "loss": 4.901210784912109,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.014810293912887573,
+ "learning_rate": 0.002552729524701833,
+ "loss": 4.895804405212402,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.010993832722306252,
+ "learning_rate": 0.0025517756472053068,
+ "loss": 4.891035079956055,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.013413915410637856,
+ "learning_rate": 0.002550820934397368,
+ "loss": 4.900317192077637,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.01580147072672844,
+ "learning_rate": 0.0025498653870472117,
+ "loss": 4.883669853210449,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01864512264728546,
+ "learning_rate": 0.0025489090059247045,
+ "loss": 4.881938457489014,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017185743898153305,
+ "learning_rate": 0.0025479517918003862,
+ "loss": 4.853325843811035,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.012045156210660934,
+ "learning_rate": 0.002546993745445466,
+ "loss": 4.9152326583862305,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.013420728035271168,
+ "learning_rate": 0.0025460348676318257,
+ "loss": 4.895130157470703,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.011561610735952854,
+ "learning_rate": 0.0025450751591320152,
+ "loss": 4.890257835388184,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013164008036255836,
+ "learning_rate": 0.0025441146207192547,
+ "loss": 4.885972023010254,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01120204571634531,
+ "learning_rate": 0.002543153253167432,
+ "loss": 4.882246971130371,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.011342627927660942,
+ "learning_rate": 0.002542191057251104,
+ "loss": 4.862927436828613,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.010852213948965073,
+ "learning_rate": 0.0025412280337454942,
+ "loss": 4.879744529724121,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.012851282022893429,
+ "learning_rate": 0.0025402641834264936,
+ "loss": 4.882238388061523,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.011839655227959156,
+ "learning_rate": 0.002539299507070659,
+ "loss": 4.85658073425293,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.011315262876451015,
+ "learning_rate": 0.002538334005455213,
+ "loss": 4.9234771728515625,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.011661105789244175,
+ "learning_rate": 0.0025373676793580416,
+ "loss": 4.855696678161621,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.009712521918118,
+ "learning_rate": 0.0025364005295576975,
+ "loss": 4.89942741394043,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.011002928018569946,
+ "learning_rate": 0.0025354325568333956,
+ "loss": 4.871459007263184,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.01509513333439827,
+ "learning_rate": 0.0025344637619650135,
+ "loss": 4.834778785705566,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 5.031810283660889,
+ "eval_runtime": 51.9412,
+ "eval_samples_per_second": 47.015,
+ "eval_steps_per_second": 1.482,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.010317740961909294,
+ "learning_rate": 0.0025334941457330925,
+ "loss": 4.8169708251953125,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.00892715621739626,
+ "learning_rate": 0.0025325237089188343,
+ "loss": 4.852178573608398,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.011608307249844074,
+ "learning_rate": 0.002531552452304103,
+ "loss": 4.790487289428711,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.013100114651024342,
+ "learning_rate": 0.0025305803766714216,
+ "loss": 4.863536834716797,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.016571057960391045,
+ "learning_rate": 0.002529607482803974,
+ "loss": 4.840987205505371,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.014595622196793556,
+ "learning_rate": 0.002528633771485604,
+ "loss": 4.815027236938477,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.014603931456804276,
+ "learning_rate": 0.002527659243500812,
+ "loss": 4.853024959564209,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.019107041880488396,
+ "learning_rate": 0.0025266838996347595,
+ "loss": 4.829808235168457,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.020292190834879875,
+ "learning_rate": 0.0025257077406732613,
+ "loss": 4.775232315063477,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.014630074612796307,
+ "learning_rate": 0.0025247307674027926,
+ "loss": 4.826682090759277,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.019507668912410736,
+ "learning_rate": 0.0025237529806104823,
+ "loss": 4.843222141265869,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.023950330913066864,
+ "learning_rate": 0.002522774381084116,
+ "loss": 4.820512771606445,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.022547820582985878,
+ "learning_rate": 0.002521794969612133,
+ "loss": 4.854910850524902,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.017819277942180634,
+ "learning_rate": 0.0025208147469836283,
+ "loss": 4.825885772705078,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.017989592626690865,
+ "learning_rate": 0.002519833713988349,
+ "loss": 4.8357014656066895,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.013516640290617943,
+ "learning_rate": 0.0025188518714166956,
+ "loss": 4.83729362487793,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.015956101939082146,
+ "learning_rate": 0.002517869220059721,
+ "loss": 4.807845592498779,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.013967473991215229,
+ "learning_rate": 0.002516885760709131,
+ "loss": 4.828944206237793,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.01270931027829647,
+ "learning_rate": 0.0025159014941572785,
+ "loss": 4.826594352722168,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.012433716095983982,
+ "learning_rate": 0.002514916421197171,
+ "loss": 4.797939777374268,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.011882808059453964,
+ "learning_rate": 0.0025139305426224636,
+ "loss": 4.78917932510376,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.011193005368113518,
+ "learning_rate": 0.0025129438592274614,
+ "loss": 4.787964820861816,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.011873407289385796,
+ "learning_rate": 0.002511956371807116,
+ "loss": 4.81754732131958,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.011407418176531792,
+ "learning_rate": 0.002510968081157029,
+ "loss": 4.781639575958252,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.009560304693877697,
+ "learning_rate": 0.002509978988073449,
+ "loss": 4.8163957595825195,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.009969786740839481,
+ "learning_rate": 0.0025089890933532697,
+ "loss": 4.795045852661133,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.010739767923951149,
+ "learning_rate": 0.0025079983977940312,
+ "loss": 4.847584247589111,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.009999158792197704,
+ "learning_rate": 0.0025070069021939194,
+ "loss": 4.834909439086914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.009302482940256596,
+ "learning_rate": 0.002506014607351765,
+ "loss": 4.808745384216309,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.010687744244933128,
+ "learning_rate": 0.0025050215140670403,
+ "loss": 4.789101600646973,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.010930616408586502,
+ "learning_rate": 0.0025040276231398637,
+ "loss": 4.817464351654053,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.013101065531373024,
+ "learning_rate": 0.0025030329353709956,
+ "loss": 4.827780723571777,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.01223307941108942,
+ "learning_rate": 0.0025020374515618373,
+ "loss": 4.8367414474487305,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.013091634958982468,
+ "learning_rate": 0.0025010411725144323,
+ "loss": 4.839454650878906,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.012114843353629112,
+ "learning_rate": 0.0025000440990314643,
+ "loss": 4.832447052001953,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014664588496088982,
+ "learning_rate": 0.0024990462319162588,
+ "loss": 4.740812301635742,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.015566283836960793,
+ "learning_rate": 0.002498047571972777,
+ "loss": 4.815671920776367,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.012440434657037258,
+ "learning_rate": 0.002497048120005623,
+ "loss": 4.769775867462158,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01175499614328146,
+ "learning_rate": 0.0024960478768200362,
+ "loss": 4.8101348876953125,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.012699508108198643,
+ "learning_rate": 0.002495046843221896,
+ "loss": 4.827281951904297,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.012438303790986538,
+ "learning_rate": 0.002494045020017715,
+ "loss": 4.786289691925049,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.01709924265742302,
+ "learning_rate": 0.002493042408014646,
+ "loss": 4.828742027282715,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.022684535011649132,
+ "learning_rate": 0.0024920390080204733,
+ "loss": 4.822258949279785,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.011984334327280521,
+ "learning_rate": 0.00249103482084362,
+ "loss": 4.792006969451904,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.013941102661192417,
+ "learning_rate": 0.00249002984729314,
+ "loss": 4.7926177978515625,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.0145634226500988,
+ "learning_rate": 0.002489024088178724,
+ "loss": 4.819155216217041,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.013391612097620964,
+ "learning_rate": 0.0024880175443106932,
+ "loss": 4.777242660522461,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01764756441116333,
+ "learning_rate": 0.0024870102165000015,
+ "loss": 4.81627893447876,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.015397441573441029,
+ "learning_rate": 0.002486002105558235,
+ "loss": 4.869495391845703,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01225620973855257,
+ "learning_rate": 0.0024849932122976105,
+ "loss": 4.784395217895508,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.01534759160131216,
+ "learning_rate": 0.002483983537530975,
+ "loss": 4.813475608825684,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.015057779848575592,
+ "learning_rate": 0.002482973082071805,
+ "loss": 4.8281569480896,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.01243986189365387,
+ "learning_rate": 0.0024819618467342066,
+ "loss": 4.7866973876953125,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.013326527550816536,
+ "learning_rate": 0.0024809498323329136,
+ "loss": 4.750214576721191,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.013644695281982422,
+ "learning_rate": 0.002479937039683288,
+ "loss": 4.790144920349121,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01296213548630476,
+ "learning_rate": 0.002478923469601319,
+ "loss": 4.812170028686523,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.011848201975226402,
+ "learning_rate": 0.0024779091229036204,
+ "loss": 4.798855781555176,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.012765968218445778,
+ "learning_rate": 0.0024768940004074335,
+ "loss": 4.8272294998168945,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.011700666509568691,
+ "learning_rate": 0.002475878102930625,
+ "loss": 4.762265205383301,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.01018498931080103,
+ "learning_rate": 0.0024748614312916843,
+ "loss": 4.796628952026367,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.010241681709885597,
+ "learning_rate": 0.002473843986309725,
+ "loss": 4.829319000244141,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.009332080371677876,
+ "learning_rate": 0.0024728257688044855,
+ "loss": 4.783773899078369,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.012860726565122604,
+ "learning_rate": 0.002471806779596324,
+ "loss": 4.802700519561768,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.011156697757542133,
+ "learning_rate": 0.0024707870195062225,
+ "loss": 4.808038234710693,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.010329720564186573,
+ "learning_rate": 0.002469766489355783,
+ "loss": 4.761645317077637,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.011093895882368088,
+ "learning_rate": 0.0024687451899672278,
+ "loss": 4.795220851898193,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.010740782134234905,
+ "learning_rate": 0.0024677231221634,
+ "loss": 4.79289436340332,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.012272508814930916,
+ "learning_rate": 0.0024667002867677606,
+ "loss": 4.811873435974121,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.008665879257023335,
+ "learning_rate": 0.0024656766846043905,
+ "loss": 4.744179725646973,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.00933373998850584,
+ "learning_rate": 0.0024646523164979866,
+ "loss": 4.805453777313232,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.010399337857961655,
+ "learning_rate": 0.002463627183273864,
+ "loss": 4.75568962097168,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.013210986740887165,
+ "learning_rate": 0.0024626012857579542,
+ "loss": 4.797974586486816,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.967026233673096,
+ "eval_runtime": 51.07,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.01731858029961586,
+ "learning_rate": 0.0024615746247768043,
+ "loss": 4.748838901519775,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.018902689218521118,
+ "learning_rate": 0.0024605472011575763,
+ "loss": 4.717338562011719,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.01734415628015995,
+ "learning_rate": 0.0024595190157280468,
+ "loss": 4.737737655639648,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.01971321739256382,
+ "learning_rate": 0.002458490069316607,
+ "loss": 4.774558067321777,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.020029844716191292,
+ "learning_rate": 0.002457460362752259,
+ "loss": 4.739579200744629,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.020853109657764435,
+ "learning_rate": 0.0024564298968646203,
+ "loss": 4.791436672210693,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.019159914925694466,
+ "learning_rate": 0.0024553986724839176,
+ "loss": 4.765683650970459,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.0165473110973835,
+ "learning_rate": 0.0024543666904409903,
+ "loss": 4.732603073120117,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.017426826059818268,
+ "learning_rate": 0.002453333951567287,
+ "loss": 4.7360639572143555,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014590734615921974,
+ "learning_rate": 0.002452300456694868,
+ "loss": 4.750031471252441,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.011172222904860973,
+ "learning_rate": 0.0024512662066563996,
+ "loss": 4.740771293640137,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.012294141575694084,
+ "learning_rate": 0.0024502312022851606,
+ "loss": 4.726214408874512,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.01146282535046339,
+ "learning_rate": 0.002449195444415034,
+ "loss": 4.753599166870117,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.011197247542440891,
+ "learning_rate": 0.0024481589338805113,
+ "loss": 4.751479148864746,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.013252820819616318,
+ "learning_rate": 0.0024471216715166904,
+ "loss": 4.743409156799316,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.012366211041808128,
+ "learning_rate": 0.0024460836581592744,
+ "loss": 4.731625556945801,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.013906655833125114,
+ "learning_rate": 0.0024450448946445737,
+ "loss": 4.768153190612793,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.012118754908442497,
+ "learning_rate": 0.002444005381809499,
+ "loss": 4.74198579788208,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.01197550818324089,
+ "learning_rate": 0.002442965120491568,
+ "loss": 4.737289905548096,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.010640372522175312,
+ "learning_rate": 0.002441924111528901,
+ "loss": 4.7472686767578125,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.013553396798670292,
+ "learning_rate": 0.0024408823557602196,
+ "loss": 4.711215972900391,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.01393711194396019,
+ "learning_rate": 0.002439839854024847,
+ "loss": 4.720060348510742,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01589774526655674,
+ "learning_rate": 0.002438796607162709,
+ "loss": 4.695697784423828,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.0097908740863204,
+ "learning_rate": 0.0024377526160143303,
+ "loss": 4.729333877563477,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.011745489202439785,
+ "learning_rate": 0.0024367078814208363,
+ "loss": 4.722957134246826,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.009649662300944328,
+ "learning_rate": 0.0024356624042239498,
+ "loss": 4.727540493011475,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.008429240435361862,
+ "learning_rate": 0.002434616185265993,
+ "loss": 4.705052375793457,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.00984937883913517,
+ "learning_rate": 0.002433569225389887,
+ "loss": 4.721161365509033,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.009237290360033512,
+ "learning_rate": 0.0024325215254391466,
+ "loss": 4.724277496337891,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.008900545537471771,
+ "learning_rate": 0.002431473086257886,
+ "loss": 4.711522102355957,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.010003026574850082,
+ "learning_rate": 0.002430423908690813,
+ "loss": 4.728816032409668,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.011658324860036373,
+ "learning_rate": 0.0024293739935832316,
+ "loss": 4.716179847717285,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.01312185451388359,
+ "learning_rate": 0.0024283233417810393,
+ "loss": 4.7172346115112305,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.014306646771728992,
+ "learning_rate": 0.0024272719541307266,
+ "loss": 4.704054832458496,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.009844634681940079,
+ "learning_rate": 0.0024262198314793783,
+ "loss": 4.72379207611084,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.00903896801173687,
+ "learning_rate": 0.0024251669746746703,
+ "loss": 4.711025238037109,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.010476433672010899,
+ "learning_rate": 0.0024241133845648703,
+ "loss": 4.711664199829102,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.009128391742706299,
+ "learning_rate": 0.0024230590619988375,
+ "loss": 4.72395133972168,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.01228832732886076,
+ "learning_rate": 0.0024220040078260195,
+ "loss": 4.747347831726074,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.011916602961719036,
+ "learning_rate": 0.002420948222896455,
+ "loss": 4.696471691131592,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.012877700850367546,
+ "learning_rate": 0.0024198917080607706,
+ "loss": 4.731119632720947,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.012967764399945736,
+ "learning_rate": 0.0024188344641701814,
+ "loss": 4.688953876495361,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.011993546038866043,
+ "learning_rate": 0.002417776492076489,
+ "loss": 4.695237159729004,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.011188571341335773,
+ "learning_rate": 0.0024167177926320833,
+ "loss": 4.699850082397461,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.010556348599493504,
+ "learning_rate": 0.0024156583666899395,
+ "loss": 4.703063011169434,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.012023619376122952,
+ "learning_rate": 0.0024145982151036176,
+ "loss": 4.6918792724609375,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01266001258045435,
+ "learning_rate": 0.0024135373387272623,
+ "loss": 4.6653289794921875,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.012152725830674171,
+ "learning_rate": 0.0024124757384156024,
+ "loss": 4.682897567749023,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.014056944288313389,
+ "learning_rate": 0.002411413415023951,
+ "loss": 4.696136474609375,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.016826556995511055,
+ "learning_rate": 0.002410350369408202,
+ "loss": 4.701657295227051,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.015048800967633724,
+ "learning_rate": 0.0024092866024248323,
+ "loss": 4.691389083862305,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.012512968853116035,
+ "learning_rate": 0.0024082221149308997,
+ "loss": 4.677886009216309,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01596425101161003,
+ "learning_rate": 0.0024071569077840425,
+ "loss": 4.718337535858154,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01640196517109871,
+ "learning_rate": 0.002406090981842479,
+ "loss": 4.697303295135498,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.012347170151770115,
+ "learning_rate": 0.002405024337965007,
+ "loss": 4.696938991546631,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01213028747588396,
+ "learning_rate": 0.0024039569770110012,
+ "loss": 4.737306594848633,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01314542256295681,
+ "learning_rate": 0.002402888899840416,
+ "loss": 4.69948673248291,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.013764189556241035,
+ "learning_rate": 0.0024018201073137805,
+ "loss": 4.722800254821777,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.011147351004183292,
+ "learning_rate": 0.002400750600292203,
+ "loss": 4.694892883300781,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.015437614172697067,
+ "learning_rate": 0.0023996803796373653,
+ "loss": 4.704752445220947,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.016402097418904305,
+ "learning_rate": 0.0023986094462115255,
+ "loss": 4.67849063873291,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.011600594967603683,
+ "learning_rate": 0.002397537800877514,
+ "loss": 4.689641952514648,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.011864298023283482,
+ "learning_rate": 0.0023964654444987373,
+ "loss": 4.670377731323242,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01409872155636549,
+ "learning_rate": 0.002395392377939173,
+ "loss": 4.696022033691406,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.012106554582715034,
+ "learning_rate": 0.0023943186020633717,
+ "loss": 4.705615043640137,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.012310667894780636,
+ "learning_rate": 0.002393244117736455,
+ "loss": 4.702986717224121,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.013501073233783245,
+ "learning_rate": 0.002392168925824116,
+ "loss": 4.6700215339660645,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.013565126806497574,
+ "learning_rate": 0.0023910930271926166,
+ "loss": 4.699914455413818,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.013421077281236649,
+ "learning_rate": 0.00239001642270879,
+ "loss": 4.713191986083984,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.014408102259039879,
+ "learning_rate": 0.002388939113240035,
+ "loss": 4.655015468597412,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.013502835296094418,
+ "learning_rate": 0.0023878610996543224,
+ "loss": 4.6142258644104,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.014169570058584213,
+ "learning_rate": 0.002386782382820188,
+ "loss": 4.660243034362793,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.885881423950195,
+ "eval_runtime": 52.0315,
+ "eval_samples_per_second": 46.933,
+ "eval_steps_per_second": 1.48,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.014618083834648132,
+ "learning_rate": 0.002385702963606734,
+ "loss": 4.640046119689941,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01745622046291828,
+ "learning_rate": 0.0023846228428836278,
+ "loss": 4.641966819763184,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.014995258301496506,
+ "learning_rate": 0.002383542021521106,
+ "loss": 4.60922908782959,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.013278020545840263,
+ "learning_rate": 0.002382460500389964,
+ "loss": 4.652071952819824,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.014685257337987423,
+ "learning_rate": 0.002381378280361566,
+ "loss": 4.672174453735352,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.012737024575471878,
+ "learning_rate": 0.002380295362307836,
+ "loss": 4.612064361572266,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.012992842122912407,
+ "learning_rate": 0.002379211747101261,
+ "loss": 4.611995697021484,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.010210646316409111,
+ "learning_rate": 0.0023781274356148912,
+ "loss": 4.6055707931518555,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.010265499353408813,
+ "learning_rate": 0.0023770424287223365,
+ "loss": 4.637089729309082,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.010174009948968887,
+ "learning_rate": 0.002375956727297768,
+ "loss": 4.620571136474609,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.009512446820735931,
+ "learning_rate": 0.0023748703322159143,
+ "loss": 4.644092559814453,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.009849252179265022,
+ "learning_rate": 0.0023737832443520655,
+ "loss": 4.625055313110352,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.010493188165128231,
+ "learning_rate": 0.0023726954645820674,
+ "loss": 4.644410610198975,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.022181199863553047,
+ "learning_rate": 0.002371606993782326,
+ "loss": 4.593084335327148,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.011906919069588184,
+ "learning_rate": 0.0023705178328298016,
+ "loss": 4.6452836990356445,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.011874495074152946,
+ "learning_rate": 0.0023694279826020122,
+ "loss": 4.599322319030762,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.00989708211272955,
+ "learning_rate": 0.0023683374439770295,
+ "loss": 4.607342720031738,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.009256849065423012,
+ "learning_rate": 0.0023672462178334817,
+ "loss": 4.621583461761475,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.013187132775783539,
+ "learning_rate": 0.0023661543050505496,
+ "loss": 4.614629745483398,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.013505950570106506,
+ "learning_rate": 0.002365061706507968,
+ "loss": 4.629693984985352,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.012032733298838139,
+ "learning_rate": 0.0023639684230860225,
+ "loss": 4.671209335327148,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.01278337650001049,
+ "learning_rate": 0.002362874455665553,
+ "loss": 4.648874759674072,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.01123690977692604,
+ "learning_rate": 0.0023617798051279496,
+ "loss": 4.611470699310303,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.012564311735332012,
+ "learning_rate": 0.002360684472355151,
+ "loss": 4.659026145935059,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.01206638477742672,
+ "learning_rate": 0.002359588458229649,
+ "loss": 4.642566204071045,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.010300020687282085,
+ "learning_rate": 0.0023584917636344806,
+ "loss": 4.629585266113281,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.0102265365421772,
+ "learning_rate": 0.0023573943894532336,
+ "loss": 4.609250068664551,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.012440572492778301,
+ "learning_rate": 0.002356296336570042,
+ "loss": 4.636714458465576,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.011830241419374943,
+ "learning_rate": 0.0023551976058695887,
+ "loss": 4.59100866317749,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.011482699774205685,
+ "learning_rate": 0.0023540981982370997,
+ "loss": 4.620088577270508,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.012950294651091099,
+ "learning_rate": 0.0023529981145583483,
+ "loss": 4.578657627105713,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.012197710573673248,
+ "learning_rate": 0.002351897355719652,
+ "loss": 4.620813846588135,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.014839230105280876,
+ "learning_rate": 0.002350795922607872,
+ "loss": 4.615731239318848,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016670791432261467,
+ "learning_rate": 0.002349693816110414,
+ "loss": 4.655855178833008,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.014426100067794323,
+ "learning_rate": 0.0023485910371152247,
+ "loss": 4.660849571228027,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.014040560461580753,
+ "learning_rate": 0.0023474875865107936,
+ "loss": 4.634917736053467,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018269391730427742,
+ "learning_rate": 0.0023463834651861496,
+ "loss": 4.611809253692627,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.013855630531907082,
+ "learning_rate": 0.0023452786740308647,
+ "loss": 4.615225791931152,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.011505477130413055,
+ "learning_rate": 0.0023441732139350487,
+ "loss": 4.62592887878418,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.010166114196181297,
+ "learning_rate": 0.002343067085789351,
+ "loss": 4.6219024658203125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.009253021329641342,
+ "learning_rate": 0.0023419602904849596,
+ "loss": 4.658320426940918,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01027526892721653,
+ "learning_rate": 0.002340852828913598,
+ "loss": 4.589777946472168,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.008855022490024567,
+ "learning_rate": 0.002339744701967529,
+ "loss": 4.588204383850098,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.00965463649481535,
+ "learning_rate": 0.002338635910539551,
+ "loss": 4.648177146911621,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.009073720313608646,
+ "learning_rate": 0.0023375264555229956,
+ "loss": 4.585582733154297,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.010105383582413197,
+ "learning_rate": 0.0023364163378117317,
+ "loss": 4.612174987792969,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.015307556837797165,
+ "learning_rate": 0.002335305558300161,
+ "loss": 4.591984748840332,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016716254875063896,
+ "learning_rate": 0.002334194117883219,
+ "loss": 4.659285545349121,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.013281598687171936,
+ "learning_rate": 0.0023330820174563725,
+ "loss": 4.645829200744629,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.014416513964533806,
+ "learning_rate": 0.0023319692579156204,
+ "loss": 4.584893226623535,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.016326071694493294,
+ "learning_rate": 0.002330855840157494,
+ "loss": 4.633223533630371,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.012851824052631855,
+ "learning_rate": 0.002329741765079053,
+ "loss": 4.580986499786377,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.015444506891071796,
+ "learning_rate": 0.0023286270335778876,
+ "loss": 4.615891456604004,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.01391979493200779,
+ "learning_rate": 0.0023275116465521174,
+ "loss": 4.628403186798096,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.012138686142861843,
+ "learning_rate": 0.0023263956049003895,
+ "loss": 4.6259942054748535,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.013741672039031982,
+ "learning_rate": 0.0023252789095218785,
+ "loss": 4.6056742668151855,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01103319600224495,
+ "learning_rate": 0.002324161561316285,
+ "loss": 4.621670246124268,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.010333316400647163,
+ "learning_rate": 0.0023230435611838367,
+ "loss": 4.620484352111816,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.011063606478273869,
+ "learning_rate": 0.0023219249100252864,
+ "loss": 4.64952278137207,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.009984857402741909,
+ "learning_rate": 0.0023208056087419106,
+ "loss": 4.572571754455566,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.009986856952309608,
+ "learning_rate": 0.002319685658235511,
+ "loss": 4.570366382598877,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.010905452072620392,
+ "learning_rate": 0.0023185650594084103,
+ "loss": 4.603614807128906,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.012800061143934727,
+ "learning_rate": 0.002317443813163456,
+ "loss": 4.562705039978027,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.01252608560025692,
+ "learning_rate": 0.002316321920404015,
+ "loss": 4.599016189575195,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.011305592954158783,
+ "learning_rate": 0.002315199382033976,
+ "loss": 4.644617557525635,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.010594413615763187,
+ "learning_rate": 0.002314076198957749,
+ "loss": 4.569808006286621,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.009859138168394566,
+ "learning_rate": 0.002312952372080261,
+ "loss": 4.5890398025512695,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.010982917621731758,
+ "learning_rate": 0.00231182790230696,
+ "loss": 4.606348991394043,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.010252755135297775,
+ "learning_rate": 0.00231070279054381,
+ "loss": 4.58942174911499,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01110257301479578,
+ "learning_rate": 0.0023095770376972947,
+ "loss": 4.599823474884033,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01003815233707428,
+ "learning_rate": 0.0023084506446744103,
+ "loss": 4.569343090057373,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.010517258197069168,
+ "learning_rate": 0.0023073236123826745,
+ "loss": 4.646301746368408,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.804611682891846,
+ "eval_runtime": 51.6279,
+ "eval_samples_per_second": 47.3,
+ "eval_steps_per_second": 1.491,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.011439066380262375,
+ "learning_rate": 0.0023061959417301145,
+ "loss": 4.563335418701172,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.01073157787322998,
+ "learning_rate": 0.002305067633625275,
+ "loss": 4.540966987609863,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.011391228064894676,
+ "learning_rate": 0.0023039386889772136,
+ "loss": 4.555583953857422,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.012814595364034176,
+ "learning_rate": 0.0023028091086955,
+ "loss": 4.517425537109375,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.012994988821446896,
+ "learning_rate": 0.002301678893690218,
+ "loss": 4.538477897644043,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.01460991520434618,
+ "learning_rate": 0.00230054804487196,
+ "loss": 4.571145057678223,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.013036075979471207,
+ "learning_rate": 0.0022994165631518306,
+ "loss": 4.519233703613281,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.011917057447135448,
+ "learning_rate": 0.0022982844494414444,
+ "loss": 4.5071587562561035,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.013928866013884544,
+ "learning_rate": 0.0022971517046529257,
+ "loss": 4.531422138214111,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01407038513571024,
+ "learning_rate": 0.0022960183296989053,
+ "loss": 4.511205196380615,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.019396567717194557,
+ "learning_rate": 0.0022948843254925245,
+ "loss": 4.530007362365723,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.014425835572183132,
+ "learning_rate": 0.002293749692947428,
+ "loss": 4.530138969421387,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.01600247249007225,
+ "learning_rate": 0.00229261443297777,
+ "loss": 4.5643391609191895,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.018238037824630737,
+ "learning_rate": 0.0022914785464982095,
+ "loss": 4.540779113769531,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.015427535399794579,
+ "learning_rate": 0.0022903420344239085,
+ "loss": 4.506287097930908,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.016816625371575356,
+ "learning_rate": 0.0022892048976705347,
+ "loss": 4.537919521331787,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.01999717950820923,
+ "learning_rate": 0.002288067137154259,
+ "loss": 4.558426380157471,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.020884918048977852,
+ "learning_rate": 0.002286928753791753,
+ "loss": 4.592915058135986,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.01620141603052616,
+ "learning_rate": 0.002285789748500194,
+ "loss": 4.553356170654297,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.018495604395866394,
+ "learning_rate": 0.0022846501221972567,
+ "loss": 4.573139190673828,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.0156320258975029,
+ "learning_rate": 0.0022835098758011174,
+ "loss": 4.585921764373779,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.014515317976474762,
+ "learning_rate": 0.002282369010230452,
+ "loss": 4.5532002449035645,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.014826556667685509,
+ "learning_rate": 0.0022812275264044355,
+ "loss": 4.560998916625977,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.013240831904113293,
+ "learning_rate": 0.00228008542524274,
+ "loss": 4.531787872314453,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.013814719393849373,
+ "learning_rate": 0.0022789427076655372,
+ "loss": 4.611872673034668,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.012246561236679554,
+ "learning_rate": 0.0022777993745934934,
+ "loss": 4.5565385818481445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01206218171864748,
+ "learning_rate": 0.002276655426947771,
+ "loss": 4.570852279663086,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01007329672574997,
+ "learning_rate": 0.0022755108656500275,
+ "loss": 4.553400039672852,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.010361720807850361,
+ "learning_rate": 0.0022743656916224173,
+ "loss": 4.527509689331055,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.009456256404519081,
+ "learning_rate": 0.0022732199057875846,
+ "loss": 4.520236015319824,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.009310049936175346,
+ "learning_rate": 0.002272073509068669,
+ "loss": 4.518343925476074,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.00927449855953455,
+ "learning_rate": 0.0022709265023893015,
+ "loss": 4.559690475463867,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.008658594451844692,
+ "learning_rate": 0.002269778886673605,
+ "loss": 4.513876914978027,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.008928220719099045,
+ "learning_rate": 0.0022686306628461918,
+ "loss": 4.545050621032715,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.008046358823776245,
+ "learning_rate": 0.002267481831832166,
+ "loss": 4.54866361618042,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.008318924345076084,
+ "learning_rate": 0.00226633239455712,
+ "loss": 4.546968936920166,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.008808869868516922,
+ "learning_rate": 0.0022651823519471343,
+ "loss": 4.50877571105957,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.008767383173108101,
+ "learning_rate": 0.0022640317049287774,
+ "loss": 4.57666540145874,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.009351354092359543,
+ "learning_rate": 0.0022628804544291047,
+ "loss": 4.511702537536621,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.010137057863175869,
+ "learning_rate": 0.002261728601375659,
+ "loss": 4.539676189422607,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.00882046204060316,
+ "learning_rate": 0.002260576146696466,
+ "loss": 4.533196449279785,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.009354782290756702,
+ "learning_rate": 0.0022594230913200388,
+ "loss": 4.515488624572754,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.009241201914846897,
+ "learning_rate": 0.002258269436175372,
+ "loss": 4.542706489562988,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.009252829477190971,
+ "learning_rate": 0.002257115182191947,
+ "loss": 4.592660903930664,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.011243131943047047,
+ "learning_rate": 0.0022559603302997225,
+ "loss": 4.527804851531982,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.011032823473215103,
+ "learning_rate": 0.0022548048814291437,
+ "loss": 4.540122032165527,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.010495275259017944,
+ "learning_rate": 0.0022536488365111347,
+ "loss": 4.532835006713867,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.009992090053856373,
+ "learning_rate": 0.0022524921964770994,
+ "loss": 4.521217346191406,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01141198631376028,
+ "learning_rate": 0.0022513349622589225,
+ "loss": 4.519709587097168,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.013087889179587364,
+ "learning_rate": 0.002250177134788966,
+ "loss": 4.547289848327637,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.01328477542847395,
+ "learning_rate": 0.0022490187150000706,
+ "loss": 4.527865409851074,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01033794041723013,
+ "learning_rate": 0.002247859703825555,
+ "loss": 4.54740047454834,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.010522300377488136,
+ "learning_rate": 0.002246700102199213,
+ "loss": 4.538840293884277,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.012151090428233147,
+ "learning_rate": 0.0022455399110553145,
+ "loss": 4.536865234375,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.010553102008998394,
+ "learning_rate": 0.0022443791313286054,
+ "loss": 4.5348334312438965,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.011433709412813187,
+ "learning_rate": 0.0022432177639543035,
+ "loss": 4.510526657104492,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01189691387116909,
+ "learning_rate": 0.002242055809868103,
+ "loss": 4.530807018280029,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.010031254962086678,
+ "learning_rate": 0.0022408932700061683,
+ "loss": 4.496850967407227,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.010129331611096859,
+ "learning_rate": 0.002239730145305138,
+ "loss": 4.551287651062012,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.010430586524307728,
+ "learning_rate": 0.0022385664367021185,
+ "loss": 4.542672157287598,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.011774476617574692,
+ "learning_rate": 0.002237402145134691,
+ "loss": 4.537034034729004,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.01184659544378519,
+ "learning_rate": 0.002236237271540903,
+ "loss": 4.521193504333496,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.011327247135341167,
+ "learning_rate": 0.0022350718168592724,
+ "loss": 4.49426794052124,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.010767918080091476,
+ "learning_rate": 0.0022339057820287855,
+ "loss": 4.531225204467773,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.011195424944162369,
+ "learning_rate": 0.002232739167988894,
+ "loss": 4.492671012878418,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.010394908487796783,
+ "learning_rate": 0.0022315719756795197,
+ "loss": 4.523616790771484,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.010277998633682728,
+ "learning_rate": 0.002230404206041047,
+ "loss": 4.487832069396973,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.010955783538520336,
+ "learning_rate": 0.0022292358600143275,
+ "loss": 4.526865005493164,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.010611076839268208,
+ "learning_rate": 0.002228066938540676,
+ "loss": 4.498682022094727,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.010008828714489937,
+ "learning_rate": 0.0022268974425618724,
+ "loss": 4.524019241333008,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.008973540738224983,
+ "learning_rate": 0.002225727373020157,
+ "loss": 4.512996673583984,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.011754963546991348,
+ "learning_rate": 0.002224556730858235,
+ "loss": 4.496376991271973,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.744190216064453,
+ "eval_runtime": 51.6572,
+ "eval_samples_per_second": 47.273,
+ "eval_steps_per_second": 1.491,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.012094628997147083,
+ "learning_rate": 0.0022233855170192703,
+ "loss": 4.443447113037109,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.014347990974783897,
+ "learning_rate": 0.00222221373244689,
+ "loss": 4.485690116882324,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.021286984905600548,
+ "learning_rate": 0.0022210413780851786,
+ "loss": 4.487505912780762,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.020273910835385323,
+ "learning_rate": 0.0022198684548786827,
+ "loss": 4.473119735717773,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.017642971128225327,
+ "learning_rate": 0.002218694963772403,
+ "loss": 4.440615653991699,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.02075147069990635,
+ "learning_rate": 0.0022175209057118017,
+ "loss": 4.527026176452637,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016479508951306343,
+ "learning_rate": 0.002216346281642796,
+ "loss": 4.508321762084961,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.017978593707084656,
+ "learning_rate": 0.002215171092511759,
+ "loss": 4.500985145568848,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.01739897020161152,
+ "learning_rate": 0.002213995339265519,
+ "loss": 4.474225044250488,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.015556776896119118,
+ "learning_rate": 0.0022128190228513597,
+ "loss": 4.444849014282227,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.016166578978300095,
+ "learning_rate": 0.002211642144217018,
+ "loss": 4.5026726722717285,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.01534261740744114,
+ "learning_rate": 0.0022104647043106832,
+ "loss": 4.484527587890625,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.013517106883227825,
+ "learning_rate": 0.002209286704080998,
+ "loss": 4.453907012939453,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.012562787160277367,
+ "learning_rate": 0.0022081081444770556,
+ "loss": 4.471528053283691,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.009947682730853558,
+ "learning_rate": 0.0022069290264483997,
+ "loss": 4.504153728485107,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01169571466743946,
+ "learning_rate": 0.002205749350945026,
+ "loss": 4.447345733642578,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.011394352652132511,
+ "learning_rate": 0.0022045691189173755,
+ "loss": 4.447491645812988,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.011168593540787697,
+ "learning_rate": 0.0022033883313163417,
+ "loss": 4.441639423370361,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.009503341279923916,
+ "learning_rate": 0.002202206989093263,
+ "loss": 4.461601257324219,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.0112199941650033,
+ "learning_rate": 0.002201025093199926,
+ "loss": 4.444095134735107,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.010505333542823792,
+ "learning_rate": 0.002199842644588562,
+ "loss": 4.460987091064453,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.009494813159108162,
+ "learning_rate": 0.0021986596442118496,
+ "loss": 4.460158348083496,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.01063452661037445,
+ "learning_rate": 0.00219747609302291,
+ "loss": 4.489625453948975,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.012194888666272163,
+ "learning_rate": 0.0021962919919753087,
+ "loss": 4.441605567932129,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.010613647289574146,
+ "learning_rate": 0.0021951073420230557,
+ "loss": 4.475821018218994,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.009643700905144215,
+ "learning_rate": 0.002193922144120601,
+ "loss": 4.467228889465332,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.00912613607943058,
+ "learning_rate": 0.0021927363992228373,
+ "loss": 4.455435752868652,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.009371735155582428,
+ "learning_rate": 0.0021915501082850988,
+ "loss": 4.479915142059326,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.008617875166237354,
+ "learning_rate": 0.002190363272263158,
+ "loss": 4.415642738342285,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.009571383707225323,
+ "learning_rate": 0.002189175892113227,
+ "loss": 4.439854621887207,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.00782635249197483,
+ "learning_rate": 0.002187987968791957,
+ "loss": 4.484311103820801,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.008408057503402233,
+ "learning_rate": 0.0021867995032564358,
+ "loss": 4.46577262878418,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.009641910903155804,
+ "learning_rate": 0.00218561049646419,
+ "loss": 4.437408447265625,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.008498182520270348,
+ "learning_rate": 0.0021844209493731784,
+ "loss": 4.442599296569824,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.009289798326790333,
+ "learning_rate": 0.0021832308629418,
+ "loss": 4.517641067504883,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.009360183961689472,
+ "learning_rate": 0.0021820402381288848,
+ "loss": 4.474143981933594,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.008764738216996193,
+ "learning_rate": 0.0021808490758936983,
+ "loss": 4.4376702308654785,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.009673701599240303,
+ "learning_rate": 0.0021796573771959375,
+ "loss": 4.435220718383789,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.010852573439478874,
+ "learning_rate": 0.0021784651429957334,
+ "loss": 4.465187072753906,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.011292314156889915,
+ "learning_rate": 0.0021772723742536464,
+ "loss": 4.458176136016846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01071334257721901,
+ "learning_rate": 0.0021760790719306702,
+ "loss": 4.425498962402344,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.01505239401012659,
+ "learning_rate": 0.0021748852369882265,
+ "loss": 4.463129043579102,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.011774689890444279,
+ "learning_rate": 0.0021736908703881663,
+ "loss": 4.462577819824219,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.011241118423640728,
+ "learning_rate": 0.002172495973092769,
+ "loss": 4.452583312988281,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.01072910986840725,
+ "learning_rate": 0.002171300546064742,
+ "loss": 4.466248035430908,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01021308358758688,
+ "learning_rate": 0.002170104590267219,
+ "loss": 4.485857963562012,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.012630444951355457,
+ "learning_rate": 0.00216890810666376,
+ "loss": 4.442112922668457,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.013602569699287415,
+ "learning_rate": 0.0021677110962183515,
+ "loss": 4.45171594619751,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.013808641582727432,
+ "learning_rate": 0.0021665135598954014,
+ "loss": 4.468818187713623,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.011510699987411499,
+ "learning_rate": 0.002165315498659744,
+ "loss": 4.439584255218506,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.010502424091100693,
+ "learning_rate": 0.0021641169134766353,
+ "loss": 4.478377342224121,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.009953407570719719,
+ "learning_rate": 0.0021629178053117542,
+ "loss": 4.474404335021973,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.011158930137753487,
+ "learning_rate": 0.0021617181751312,
+ "loss": 4.4549560546875,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.010431940667331219,
+ "learning_rate": 0.002160518023901493,
+ "loss": 4.444911003112793,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.009100335650146008,
+ "learning_rate": 0.0021593173525895734,
+ "loss": 4.428071975708008,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.009362847544252872,
+ "learning_rate": 0.0021581161621628013,
+ "loss": 4.462827682495117,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.009893129579722881,
+ "learning_rate": 0.0021569144535889535,
+ "loss": 4.426912307739258,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.010524468496441841,
+ "learning_rate": 0.0021557122278362244,
+ "loss": 4.435606002807617,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.010621445253491402,
+ "learning_rate": 0.002154509485873226,
+ "loss": 4.454901695251465,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.010836896486580372,
+ "learning_rate": 0.0021533062286689855,
+ "loss": 4.4441728591918945,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.011403030715882778,
+ "learning_rate": 0.0021521024571929457,
+ "loss": 4.483884811401367,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.012522426433861256,
+ "learning_rate": 0.0021508981724149643,
+ "loss": 4.464006423950195,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.014799179509282112,
+ "learning_rate": 0.0021496933753053105,
+ "loss": 4.460402011871338,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.012453878298401833,
+ "learning_rate": 0.0021484880668346685,
+ "loss": 4.456899642944336,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.012270379811525345,
+ "learning_rate": 0.002147282247974133,
+ "loss": 4.477325439453125,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01094560045748949,
+ "learning_rate": 0.0021460759196952105,
+ "loss": 4.422362327575684,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.009995176456868649,
+ "learning_rate": 0.0021448690829698183,
+ "loss": 4.405127048492432,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.014935536310076714,
+ "learning_rate": 0.002143661738770283,
+ "loss": 4.445840358734131,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.011492030695080757,
+ "learning_rate": 0.0021424538880693393,
+ "loss": 4.459555625915527,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.008793608285486698,
+ "learning_rate": 0.002141245531840132,
+ "loss": 4.436631202697754,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.008894025348126888,
+ "learning_rate": 0.0021400366710562103,
+ "loss": 4.473924160003662,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.010324863716959953,
+ "learning_rate": 0.0021388273066915327,
+ "loss": 4.475680351257324,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.688324451446533,
+ "eval_runtime": 54.4515,
+ "eval_samples_per_second": 44.847,
+ "eval_steps_per_second": 1.414,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.009235628880560398,
+ "learning_rate": 0.0021376174397204615,
+ "loss": 4.370064735412598,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.008818200789391994,
+ "learning_rate": 0.0021364070711177643,
+ "loss": 4.371678352355957,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.010333757847547531,
+ "learning_rate": 0.0021351962018586147,
+ "loss": 4.3896379470825195,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.011938809417188168,
+ "learning_rate": 0.002133984832918587,
+ "loss": 4.361343860626221,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.010634614154696465,
+ "learning_rate": 0.00213277296527366,
+ "loss": 4.362449645996094,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.008545687422156334,
+ "learning_rate": 0.0021315605999002126,
+ "loss": 4.3685503005981445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.009159795008599758,
+ "learning_rate": 0.002130347737775027,
+ "loss": 4.379969596862793,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.010464153252542019,
+ "learning_rate": 0.0021291343798752835,
+ "loss": 4.376113414764404,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.011010527610778809,
+ "learning_rate": 0.0021279205271785622,
+ "loss": 4.390711784362793,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.010350867174565792,
+ "learning_rate": 0.002126706180662843,
+ "loss": 4.399683952331543,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.009708237834274769,
+ "learning_rate": 0.002125491341306503,
+ "loss": 4.363435745239258,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.010087115690112114,
+ "learning_rate": 0.002124276010088316,
+ "loss": 4.366623878479004,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.009712349623441696,
+ "learning_rate": 0.002123060187987453,
+ "loss": 4.360814571380615,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.010174986906349659,
+ "learning_rate": 0.0021218438759834785,
+ "loss": 4.347475051879883,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.010195785202085972,
+ "learning_rate": 0.0021206270750563543,
+ "loss": 4.399251461029053,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.010786959901452065,
+ "learning_rate": 0.0021194097861864354,
+ "loss": 4.397933006286621,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.011476189829409122,
+ "learning_rate": 0.002118192010354468,
+ "loss": 4.3748369216918945,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.011571013368666172,
+ "learning_rate": 0.0021169737485415927,
+ "loss": 4.388349533081055,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.011140935122966766,
+ "learning_rate": 0.002115755001729341,
+ "loss": 4.398018836975098,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.013216734863817692,
+ "learning_rate": 0.0021145357708996365,
+ "loss": 4.406947135925293,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.013385126367211342,
+ "learning_rate": 0.0021133160570347893,
+ "loss": 4.41457462310791,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.011784231290221214,
+ "learning_rate": 0.002112095861117502,
+ "loss": 4.403835296630859,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.009813498705625534,
+ "learning_rate": 0.0021108751841308645,
+ "loss": 4.3865647315979,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.010297936387360096,
+ "learning_rate": 0.0021096540270583533,
+ "loss": 4.395583152770996,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.011965147219598293,
+ "learning_rate": 0.002108432390883834,
+ "loss": 4.385443687438965,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.012212193571031094,
+ "learning_rate": 0.0021072102765915556,
+ "loss": 4.422321319580078,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.009840975515544415,
+ "learning_rate": 0.0021059876851661544,
+ "loss": 4.3568902015686035,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.013039806857705116,
+ "learning_rate": 0.002104764617592649,
+ "loss": 4.314867973327637,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.015091902576386929,
+ "learning_rate": 0.0021035410748564444,
+ "loss": 4.392338752746582,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.015448705293238163,
+ "learning_rate": 0.0021023170579433257,
+ "loss": 4.379609107971191,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.018393104895949364,
+ "learning_rate": 0.0021010925678394622,
+ "loss": 4.418291091918945,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.01866372860968113,
+ "learning_rate": 0.002099867605531402,
+ "loss": 4.4006452560424805,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.01819106750190258,
+ "learning_rate": 0.002098642172006077,
+ "loss": 4.433975696563721,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.019774602726101875,
+ "learning_rate": 0.0020974162682507952,
+ "loss": 4.417821884155273,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.016885802149772644,
+ "learning_rate": 0.0020961898952532457,
+ "loss": 4.397073268890381,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.015750881284475327,
+ "learning_rate": 0.0020949630540014956,
+ "loss": 4.4356608390808105,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.015696514397859573,
+ "learning_rate": 0.002093735745483988,
+ "loss": 4.422449588775635,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.017195021733641624,
+ "learning_rate": 0.0020925079706895435,
+ "loss": 4.443209648132324,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.015864048153162003,
+ "learning_rate": 0.0020912797306073577,
+ "loss": 4.4340691566467285,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.009857800789177418,
+ "learning_rate": 0.0020900510262270022,
+ "loss": 4.37935209274292,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.013739476911723614,
+ "learning_rate": 0.0020888218585384216,
+ "loss": 4.412148475646973,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.011667690239846706,
+ "learning_rate": 0.0020875922285319344,
+ "loss": 4.391700744628906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.012323460541665554,
+ "learning_rate": 0.0020863621371982302,
+ "loss": 4.423880577087402,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.010353626683354378,
+ "learning_rate": 0.0020851315855283724,
+ "loss": 4.427382469177246,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.0101751908659935,
+ "learning_rate": 0.002083900574513794,
+ "loss": 4.415317535400391,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.01126549206674099,
+ "learning_rate": 0.002082669105146298,
+ "loss": 4.402498245239258,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.011096738278865814,
+ "learning_rate": 0.0020814371784180575,
+ "loss": 4.433567047119141,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.010944445617496967,
+ "learning_rate": 0.0020802047953216137,
+ "loss": 4.406990051269531,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.008965318091213703,
+ "learning_rate": 0.0020789719568498747,
+ "loss": 4.3721771240234375,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.00828113779425621,
+ "learning_rate": 0.002077738663996117,
+ "loss": 4.389828205108643,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.008599454537034035,
+ "learning_rate": 0.002076504917753982,
+ "loss": 4.411396026611328,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.007713889703154564,
+ "learning_rate": 0.0020752707191174762,
+ "loss": 4.355343341827393,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.00912355538457632,
+ "learning_rate": 0.0020740360690809723,
+ "loss": 4.423413276672363,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.008269775658845901,
+ "learning_rate": 0.0020728009686392045,
+ "loss": 4.390920639038086,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.007461601868271828,
+ "learning_rate": 0.002071565418787272,
+ "loss": 4.365501403808594,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.007951333187520504,
+ "learning_rate": 0.0020703294205206336,
+ "loss": 4.378884315490723,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.007816503755748272,
+ "learning_rate": 0.002069092974835112,
+ "loss": 4.4241743087768555,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.00833756010979414,
+ "learning_rate": 0.0020678560827268886,
+ "loss": 4.391102313995361,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.008241684176027775,
+ "learning_rate": 0.0020666187451925055,
+ "loss": 4.374970436096191,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.008245996199548244,
+ "learning_rate": 0.002065380963228862,
+ "loss": 4.380222320556641,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.007378706708550453,
+ "learning_rate": 0.0020641427378332183,
+ "loss": 4.391451358795166,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.007594958413392305,
+ "learning_rate": 0.002062904070003189,
+ "loss": 4.372774124145508,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.0076752156019210815,
+ "learning_rate": 0.002061664960736747,
+ "loss": 4.384668827056885,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.008727089501917362,
+ "learning_rate": 0.00206042541103222,
+ "loss": 4.363401412963867,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.007616396062076092,
+ "learning_rate": 0.002059185421888291,
+ "loss": 4.413188934326172,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.008922954089939594,
+ "learning_rate": 0.0020579449943039967,
+ "loss": 4.383417129516602,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.009063739329576492,
+ "learning_rate": 0.002056704129278727,
+ "loss": 4.385283470153809,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.008912574499845505,
+ "learning_rate": 0.002055462827812225,
+ "loss": 4.378478050231934,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.008317525498569012,
+ "learning_rate": 0.0020542210909045844,
+ "loss": 4.367859840393066,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.009347227402031422,
+ "learning_rate": 0.0020529789195562506,
+ "loss": 4.38238000869751,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.011030065827071667,
+ "learning_rate": 0.0020517363147680183,
+ "loss": 4.372624397277832,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.011519700288772583,
+ "learning_rate": 0.002050493277541032,
+ "loss": 4.3906121253967285,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.641542434692383,
+ "eval_runtime": 51.6342,
+ "eval_samples_per_second": 47.294,
+ "eval_steps_per_second": 1.491,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.009694516658782959,
+ "learning_rate": 0.0020492498088767836,
+ "loss": 4.32808256149292,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.0097952950745821,
+ "learning_rate": 0.002048005909777114,
+ "loss": 4.280096054077148,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.010393359698355198,
+ "learning_rate": 0.0020467615812442106,
+ "loss": 4.317824363708496,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.00977372657507658,
+ "learning_rate": 0.0020455168242806054,
+ "loss": 4.3445329666137695,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.008568719029426575,
+ "learning_rate": 0.0020442716398891775,
+ "loss": 4.318026542663574,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.009732510894536972,
+ "learning_rate": 0.0020430260290731487,
+ "loss": 4.330609321594238,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.010429969057440758,
+ "learning_rate": 0.0020417799928360864,
+ "loss": 4.316794395446777,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.009292383678257465,
+ "learning_rate": 0.002040533532181898,
+ "loss": 4.317420959472656,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.011222455650568008,
+ "learning_rate": 0.0020392866481148356,
+ "loss": 4.292610168457031,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.013591725379228592,
+ "learning_rate": 0.0020380393416394894,
+ "loss": 4.325689315795898,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.013572026044130325,
+ "learning_rate": 0.002036791613760794,
+ "loss": 4.334960460662842,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.012986521236598492,
+ "learning_rate": 0.0020355434654840197,
+ "loss": 4.33829402923584,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.013321424834430218,
+ "learning_rate": 0.0020342948978147777,
+ "loss": 4.2926926612854,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.012307215481996536,
+ "learning_rate": 0.0020330459117590153,
+ "loss": 4.32757568359375,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.012909360229969025,
+ "learning_rate": 0.0020317965083230197,
+ "loss": 4.362889289855957,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017746610566973686,
+ "learning_rate": 0.0020305466885134115,
+ "loss": 4.3426313400268555,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.01696564070880413,
+ "learning_rate": 0.002029296453337148,
+ "loss": 4.3298516273498535,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.013348255306482315,
+ "learning_rate": 0.002028045803801522,
+ "loss": 4.360220432281494,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.013225145637989044,
+ "learning_rate": 0.002026794740914158,
+ "loss": 4.346250057220459,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.01569020375609398,
+ "learning_rate": 0.002025543265683016,
+ "loss": 4.308025360107422,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.018612321466207504,
+ "learning_rate": 0.002024291379116385,
+ "loss": 4.330190181732178,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.01619943603873253,
+ "learning_rate": 0.00202303908222289,
+ "loss": 4.339552402496338,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.015385213308036327,
+ "learning_rate": 0.0020217863760114826,
+ "loss": 4.3412184715271,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.018517112359404564,
+ "learning_rate": 0.002020533261491446,
+ "loss": 4.3289313316345215,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.014868509024381638,
+ "learning_rate": 0.0020192797396723903,
+ "loss": 4.359846115112305,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.014240248128771782,
+ "learning_rate": 0.0020180258115642578,
+ "loss": 4.369707107543945,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.018113158643245697,
+ "learning_rate": 0.0020167714781773143,
+ "loss": 4.362668514251709,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.01802460104227066,
+ "learning_rate": 0.002015516740522154,
+ "loss": 4.344742298126221,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.01414549071341753,
+ "learning_rate": 0.0020142615996096964,
+ "loss": 4.358080863952637,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.012157700955867767,
+ "learning_rate": 0.002013006056451186,
+ "loss": 4.347763538360596,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.011268534697592258,
+ "learning_rate": 0.0020117501120581905,
+ "loss": 4.335859298706055,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.010395628400146961,
+ "learning_rate": 0.002010493767442602,
+ "loss": 4.341609001159668,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.012147950939834118,
+ "learning_rate": 0.002009237023616635,
+ "loss": 4.328502655029297,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.00945813488215208,
+ "learning_rate": 0.0020079798815928252,
+ "loss": 4.311186790466309,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.009489500895142555,
+ "learning_rate": 0.0020067223423840294,
+ "loss": 4.371470928192139,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.009360029362142086,
+ "learning_rate": 0.0020054644070034234,
+ "loss": 4.33941125869751,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.010032779537141323,
+ "learning_rate": 0.0020042060764645037,
+ "loss": 4.365797519683838,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.009367046877741814,
+ "learning_rate": 0.002002947351781084,
+ "loss": 4.29443359375,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.008369733579456806,
+ "learning_rate": 0.0020016882339672967,
+ "loss": 4.294233322143555,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.010394785553216934,
+ "learning_rate": 0.0020004287240375893,
+ "loss": 4.352355003356934,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01068816613405943,
+ "learning_rate": 0.001999168823006728,
+ "loss": 4.342516899108887,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.009975973516702652,
+ "learning_rate": 0.0019979085318897905,
+ "loss": 4.346763610839844,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.009618151932954788,
+ "learning_rate": 0.001996647851702171,
+ "loss": 4.329115867614746,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.008399694226682186,
+ "learning_rate": 0.001995386783459577,
+ "loss": 4.28736686706543,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.008235268294811249,
+ "learning_rate": 0.001994125328178029,
+ "loss": 4.339872360229492,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.00836547277867794,
+ "learning_rate": 0.0019928634868738577,
+ "loss": 4.328156471252441,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.009619626216590405,
+ "learning_rate": 0.001991601260563706,
+ "loss": 4.337394714355469,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.009814038872718811,
+ "learning_rate": 0.0019903386502645275,
+ "loss": 4.325926780700684,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.009588954970240593,
+ "learning_rate": 0.001989075656993584,
+ "loss": 4.286477088928223,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.010471943765878677,
+ "learning_rate": 0.001987812281768447,
+ "loss": 4.322466850280762,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.010239035822451115,
+ "learning_rate": 0.001986548525606994,
+ "loss": 4.3487653732299805,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.008935568854212761,
+ "learning_rate": 0.001985284389527411,
+ "loss": 4.303601264953613,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.008624759502708912,
+ "learning_rate": 0.00198401987454819,
+ "loss": 4.34036922454834,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.009296547621488571,
+ "learning_rate": 0.0019827549816881277,
+ "loss": 4.273357391357422,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.009071138687431812,
+ "learning_rate": 0.0019814897119663248,
+ "loss": 4.321051597595215,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.008341389708220959,
+ "learning_rate": 0.0019802240664021876,
+ "loss": 4.343242645263672,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.007881715893745422,
+ "learning_rate": 0.0019789580460154227,
+ "loss": 4.32067346572876,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.008750553242862225,
+ "learning_rate": 0.0019776916518260404,
+ "loss": 4.330263614654541,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.008591460064053535,
+ "learning_rate": 0.001976424884854351,
+ "loss": 4.311957359313965,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.008084543980658054,
+ "learning_rate": 0.001975157746120966,
+ "loss": 4.331319808959961,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.008036594837903976,
+ "learning_rate": 0.001973890236646797,
+ "loss": 4.382656097412109,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.00805772002786398,
+ "learning_rate": 0.0019726223574530527,
+ "loss": 4.297247886657715,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.009019420482218266,
+ "learning_rate": 0.0019713541095612405,
+ "loss": 4.326526165008545,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.009953883476555347,
+ "learning_rate": 0.0019700854939931653,
+ "loss": 4.327356338500977,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.009340492077171803,
+ "learning_rate": 0.0019688165117709273,
+ "loss": 4.343993186950684,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.010395721532404423,
+ "learning_rate": 0.0019675471639169226,
+ "loss": 4.284934997558594,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.009946230798959732,
+ "learning_rate": 0.0019662774514538415,
+ "loss": 4.288936614990234,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.010981100611388683,
+ "learning_rate": 0.0019650073754046695,
+ "loss": 4.339935779571533,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.009688783437013626,
+ "learning_rate": 0.0019637369367926827,
+ "loss": 4.381425857543945,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.009442255832254887,
+ "learning_rate": 0.001962466136641452,
+ "loss": 4.304185390472412,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.009371291846036911,
+ "learning_rate": 0.0019611949759748364,
+ "loss": 4.306120872497559,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.009544814005494118,
+ "learning_rate": 0.001959923455816988,
+ "loss": 4.326251983642578,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.602745532989502,
+ "eval_runtime": 53.5041,
+ "eval_samples_per_second": 45.641,
+ "eval_steps_per_second": 1.439,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.01010364294052124,
+ "learning_rate": 0.0019586515771923474,
+ "loss": 4.24491024017334,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.012722729705274105,
+ "learning_rate": 0.0019573793411256446,
+ "loss": 4.257739543914795,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.01307204831391573,
+ "learning_rate": 0.0019561067486418967,
+ "loss": 4.291684150695801,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.011215966194868088,
+ "learning_rate": 0.0019548338007664098,
+ "loss": 4.237947463989258,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.012943019159138203,
+ "learning_rate": 0.0019535604985247727,
+ "loss": 4.242859363555908,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.01337282732129097,
+ "learning_rate": 0.0019522868429428636,
+ "loss": 4.238934516906738,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.013697991147637367,
+ "learning_rate": 0.0019510128350468438,
+ "loss": 4.285760402679443,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.012182159349322319,
+ "learning_rate": 0.001949738475863158,
+ "loss": 4.290044784545898,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.012620339170098305,
+ "learning_rate": 0.0019484637664185346,
+ "loss": 4.267187118530273,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.011284958571195602,
+ "learning_rate": 0.0019471887077399837,
+ "loss": 4.26662540435791,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.01161450706422329,
+ "learning_rate": 0.001945913300854797,
+ "loss": 4.283192157745361,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.010022968985140324,
+ "learning_rate": 0.001944637546790548,
+ "loss": 4.295843124389648,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.010341878980398178,
+ "learning_rate": 0.001943361446575087,
+ "loss": 4.295636177062988,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.011549185030162334,
+ "learning_rate": 0.0019420850012365457,
+ "loss": 4.288768768310547,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.012518810108304024,
+ "learning_rate": 0.0019408082118033331,
+ "loss": 4.229338645935059,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.012035328894853592,
+ "learning_rate": 0.0019395310793041353,
+ "loss": 4.258639335632324,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.011193782091140747,
+ "learning_rate": 0.0019382536047679146,
+ "loss": 4.239264488220215,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.009930136613547802,
+ "learning_rate": 0.0019369757892239096,
+ "loss": 4.252286434173584,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.00953229982405901,
+ "learning_rate": 0.0019356976337016327,
+ "loss": 4.293435096740723,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.010618340224027634,
+ "learning_rate": 0.0019344191392308716,
+ "loss": 4.235588550567627,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.008806651458144188,
+ "learning_rate": 0.0019331403068416856,
+ "loss": 4.26716423034668,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.008934796787798405,
+ "learning_rate": 0.0019318611375644065,
+ "loss": 4.258462905883789,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.009283283725380898,
+ "learning_rate": 0.0019305816324296386,
+ "loss": 4.257868766784668,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.009224812500178814,
+ "learning_rate": 0.0019293017924682555,
+ "loss": 4.258486747741699,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.008957969956099987,
+ "learning_rate": 0.0019280216187114016,
+ "loss": 4.309563636779785,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.010284051299095154,
+ "learning_rate": 0.0019267411121904894,
+ "loss": 4.275271415710449,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.009167762473225594,
+ "learning_rate": 0.0019254602739372001,
+ "loss": 4.238452911376953,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.009411636739969254,
+ "learning_rate": 0.0019241791049834816,
+ "loss": 4.2735595703125,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.009044171310961246,
+ "learning_rate": 0.0019228976063615497,
+ "loss": 4.27017879486084,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.008840068243443966,
+ "learning_rate": 0.0019216157791038834,
+ "loss": 4.28773307800293,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.009512359276413918,
+ "learning_rate": 0.001920333624243228,
+ "loss": 4.266008377075195,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.009946667589247227,
+ "learning_rate": 0.001919051142812593,
+ "loss": 4.279062271118164,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.00899894442409277,
+ "learning_rate": 0.0019177683358452505,
+ "loss": 4.277679443359375,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.00909668579697609,
+ "learning_rate": 0.0019164852043747338,
+ "loss": 4.311397552490234,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.009787327609956264,
+ "learning_rate": 0.0019152017494348408,
+ "loss": 4.30875825881958,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.008983899839222431,
+ "learning_rate": 0.001913917972059626,
+ "loss": 4.259649276733398,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.008411582559347153,
+ "learning_rate": 0.0019126338732834071,
+ "loss": 4.259811878204346,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.009162918664515018,
+ "learning_rate": 0.0019113494541407579,
+ "loss": 4.271515846252441,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.011304561980068684,
+ "learning_rate": 0.0019100647156665125,
+ "loss": 4.269217491149902,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.012483394704759121,
+ "learning_rate": 0.0019087796588957617,
+ "loss": 4.2949299812316895,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.012091380544006824,
+ "learning_rate": 0.0019074942848638522,
+ "loss": 4.289752006530762,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.010011567734181881,
+ "learning_rate": 0.0019062085946063869,
+ "loss": 4.294095993041992,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.008766217157244682,
+ "learning_rate": 0.0019049225891592226,
+ "loss": 4.256425857543945,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.012862488627433777,
+ "learning_rate": 0.001903636269558471,
+ "loss": 4.287847995758057,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.013917063362896442,
+ "learning_rate": 0.0019023496368404972,
+ "loss": 4.282784938812256,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.012129023671150208,
+ "learning_rate": 0.0019010626920419173,
+ "loss": 4.2486419677734375,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.010348426178097725,
+ "learning_rate": 0.0018997754361995986,
+ "loss": 4.289836883544922,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.011821693740785122,
+ "learning_rate": 0.001898487870350662,
+ "loss": 4.214344024658203,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.013136724010109901,
+ "learning_rate": 0.0018971999955324742,
+ "loss": 4.24208402633667,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.013173168525099754,
+ "learning_rate": 0.0018959118127826538,
+ "loss": 4.270164489746094,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.011018159799277782,
+ "learning_rate": 0.0018946233231390656,
+ "loss": 4.2571916580200195,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.01018658559769392,
+ "learning_rate": 0.0018933345276398227,
+ "loss": 4.283267498016357,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.009412688203155994,
+ "learning_rate": 0.0018920454273232848,
+ "loss": 4.272198677062988,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.007799267303198576,
+ "learning_rate": 0.0018907560232280557,
+ "loss": 4.260125637054443,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.008747464045882225,
+ "learning_rate": 0.0018894663163929863,
+ "loss": 4.2539215087890625,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.008003434166312218,
+ "learning_rate": 0.0018881763078571689,
+ "loss": 4.245024681091309,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.007593846879899502,
+ "learning_rate": 0.0018868859986599415,
+ "loss": 4.237929344177246,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.008099420927464962,
+ "learning_rate": 0.0018855953898408815,
+ "loss": 4.2789106369018555,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.008270486257970333,
+ "learning_rate": 0.0018843044824398097,
+ "loss": 4.263522148132324,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.009260379709303379,
+ "learning_rate": 0.0018830132774967866,
+ "loss": 4.275835037231445,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.009289492852985859,
+ "learning_rate": 0.0018817217760521132,
+ "loss": 4.313778400421143,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.009493921883404255,
+ "learning_rate": 0.0018804299791463281,
+ "loss": 4.303672790527344,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.009020023047924042,
+ "learning_rate": 0.0018791378878202105,
+ "loss": 4.245473384857178,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.009446556679904461,
+ "learning_rate": 0.0018778455031147724,
+ "loss": 4.267327785491943,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.009499974548816681,
+ "learning_rate": 0.0018765528260712672,
+ "loss": 4.301205635070801,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.008399393409490585,
+ "learning_rate": 0.00187525985773118,
+ "loss": 4.307802677154541,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.009519701823592186,
+ "learning_rate": 0.0018739665991362327,
+ "loss": 4.284796714782715,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.01125232595950365,
+ "learning_rate": 0.0018726730513283804,
+ "loss": 4.285930633544922,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.011608537286520004,
+ "learning_rate": 0.00187137921534981,
+ "loss": 4.257689476013184,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.01064533181488514,
+ "learning_rate": 0.001870085092242944,
+ "loss": 4.249063968658447,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.009854599833488464,
+ "learning_rate": 0.0018687906830504321,
+ "loss": 4.255159378051758,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.01121739111840725,
+ "learning_rate": 0.0018674959888151576,
+ "loss": 4.287777900695801,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.5675368309021,
+ "eval_runtime": 51.8555,
+ "eval_samples_per_second": 47.092,
+ "eval_steps_per_second": 1.485,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.0104464590549469,
+ "learning_rate": 0.0018662010105802314,
+ "loss": 4.206080436706543,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.012800181284546852,
+ "learning_rate": 0.0018649057493889945,
+ "loss": 4.213550567626953,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.013382901437580585,
+ "learning_rate": 0.001863610206285015,
+ "loss": 4.190858840942383,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.014347050338983536,
+ "learning_rate": 0.0018623143823120894,
+ "loss": 4.22884464263916,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.01347152329981327,
+ "learning_rate": 0.001861018278514239,
+ "loss": 4.215579032897949,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.012786377221345901,
+ "learning_rate": 0.0018597218959357112,
+ "loss": 4.174999237060547,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.014137202873826027,
+ "learning_rate": 0.001858425235620979,
+ "loss": 4.21709680557251,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.013864399865269661,
+ "learning_rate": 0.001857128298614737,
+ "loss": 4.2375030517578125,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.010397658683359623,
+ "learning_rate": 0.0018558310859619048,
+ "loss": 4.231938362121582,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.00888819433748722,
+ "learning_rate": 0.0018545335987076223,
+ "loss": 4.195562839508057,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.010698067955672741,
+ "learning_rate": 0.001853235837897252,
+ "loss": 4.22847318649292,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.00987589918076992,
+ "learning_rate": 0.0018519378045763766,
+ "loss": 4.2147040367126465,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.011451782658696175,
+ "learning_rate": 0.001850639499790798,
+ "loss": 4.179263114929199,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.011224704794585705,
+ "learning_rate": 0.0018493409245865368,
+ "loss": 4.23030948638916,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.00996861606836319,
+ "learning_rate": 0.0018480420800098313,
+ "loss": 4.194283962249756,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.012143119238317013,
+ "learning_rate": 0.0018467429671071374,
+ "loss": 4.221704959869385,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.013126783072948456,
+ "learning_rate": 0.0018454435869251268,
+ "loss": 4.187748908996582,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.013031953014433384,
+ "learning_rate": 0.001844143940510687,
+ "loss": 4.248250961303711,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.012868993915617466,
+ "learning_rate": 0.001842844028910919,
+ "loss": 4.20161247253418,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.010102041065692902,
+ "learning_rate": 0.0018415438531731395,
+ "loss": 4.213654518127441,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.009599820710718632,
+ "learning_rate": 0.0018402434143448747,
+ "loss": 4.206313133239746,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.011536866426467896,
+ "learning_rate": 0.0018389427134738658,
+ "loss": 4.237072944641113,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.009558089077472687,
+ "learning_rate": 0.0018376417516080636,
+ "loss": 4.182195663452148,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01056219544261694,
+ "learning_rate": 0.0018363405297956302,
+ "loss": 4.232585906982422,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.012673584744334221,
+ "learning_rate": 0.0018350390490849354,
+ "loss": 4.228830337524414,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.011215884238481522,
+ "learning_rate": 0.0018337373105245598,
+ "loss": 4.216056823730469,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.011281915940344334,
+ "learning_rate": 0.00183243531516329,
+ "loss": 4.233404159545898,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.010381227359175682,
+ "learning_rate": 0.0018311330640501204,
+ "loss": 4.182013034820557,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.00900198519229889,
+ "learning_rate": 0.0018298305582342513,
+ "loss": 4.215488433837891,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.010198601521551609,
+ "learning_rate": 0.0018285277987650876,
+ "loss": 4.19389533996582,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.009965402074158192,
+ "learning_rate": 0.0018272247866922401,
+ "loss": 4.218005657196045,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.009299935773015022,
+ "learning_rate": 0.0018259215230655202,
+ "loss": 4.223684310913086,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.008977951481938362,
+ "learning_rate": 0.0018246180089349462,
+ "loss": 4.215778350830078,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.009261307306587696,
+ "learning_rate": 0.0018233142453507344,
+ "loss": 4.215066909790039,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.008322088047862053,
+ "learning_rate": 0.0018220102333633044,
+ "loss": 4.2188005447387695,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.008716144599020481,
+ "learning_rate": 0.0018207059740232739,
+ "loss": 4.187867641448975,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.008897020481526852,
+ "learning_rate": 0.0018194014683814625,
+ "loss": 4.218034267425537,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.007298174314200878,
+ "learning_rate": 0.0018180967174888857,
+ "loss": 4.2017011642456055,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.008491789922118187,
+ "learning_rate": 0.0018167917223967582,
+ "loss": 4.239873886108398,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.008413378149271011,
+ "learning_rate": 0.0018154864841564908,
+ "loss": 4.234771251678467,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.008141471073031425,
+ "learning_rate": 0.0018141810038196909,
+ "loss": 4.185944080352783,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.009480944834649563,
+ "learning_rate": 0.0018128752824381603,
+ "loss": 4.206586837768555,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.0106047997251153,
+ "learning_rate": 0.001811569321063894,
+ "loss": 4.235287666320801,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.010663514956831932,
+ "learning_rate": 0.0018102631207490838,
+ "loss": 4.204899311065674,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.011473523452877998,
+ "learning_rate": 0.001808956682546109,
+ "loss": 4.253668308258057,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01165992021560669,
+ "learning_rate": 0.0018076500075075456,
+ "loss": 4.209845542907715,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.011692410334944725,
+ "learning_rate": 0.0018063430966861564,
+ "loss": 4.250368595123291,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01061506848782301,
+ "learning_rate": 0.0018050359511348965,
+ "loss": 4.193821907043457,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.010510941036045551,
+ "learning_rate": 0.0018037285719069097,
+ "loss": 4.254327774047852,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.011587041430175304,
+ "learning_rate": 0.001802420960055528,
+ "loss": 4.2838945388793945,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.01257420051842928,
+ "learning_rate": 0.0018011131166342693,
+ "loss": 4.258241653442383,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.00936233438551426,
+ "learning_rate": 0.0017998050426968407,
+ "loss": 4.201854705810547,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.010093127377331257,
+ "learning_rate": 0.0017984967392971338,
+ "loss": 4.193051338195801,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.010032009333372116,
+ "learning_rate": 0.0017971882074892236,
+ "loss": 4.230212211608887,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.009455587714910507,
+ "learning_rate": 0.0017958794483273726,
+ "loss": 4.2157673835754395,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.009549238719046116,
+ "learning_rate": 0.001794570462866023,
+ "loss": 4.260272026062012,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.011059699580073357,
+ "learning_rate": 0.0017932612521598013,
+ "loss": 4.214818000793457,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.011653627268970013,
+ "learning_rate": 0.0017919518172635146,
+ "loss": 4.205112457275391,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.012438314035534859,
+ "learning_rate": 0.001790642159232151,
+ "loss": 4.179045677185059,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.011234673671424389,
+ "learning_rate": 0.0017893322791208784,
+ "loss": 4.239317417144775,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.010923845693469048,
+ "learning_rate": 0.0017880221779850434,
+ "loss": 4.230571269989014,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.009219309315085411,
+ "learning_rate": 0.001786711856880171,
+ "loss": 4.229068756103516,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.009849398396909237,
+ "learning_rate": 0.0017854013168619638,
+ "loss": 4.254772186279297,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.01117763202637434,
+ "learning_rate": 0.001784090558986299,
+ "loss": 4.252647399902344,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.010664612054824829,
+ "learning_rate": 0.0017827795843092316,
+ "loss": 4.204596519470215,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.00847950391471386,
+ "learning_rate": 0.0017814683938869895,
+ "loss": 4.211061954498291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.007305726408958435,
+ "learning_rate": 0.0017801569887759753,
+ "loss": 4.181221008300781,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.007970666512846947,
+ "learning_rate": 0.0017788453700327653,
+ "loss": 4.237855434417725,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.008873614482581615,
+ "learning_rate": 0.0017775335387141054,
+ "loss": 4.229188919067383,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.009691286832094193,
+ "learning_rate": 0.0017762214958769156,
+ "loss": 4.205491542816162,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.009360672906041145,
+ "learning_rate": 0.0017749092425782852,
+ "loss": 4.233466625213623,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.009895884431898594,
+ "learning_rate": 0.0017735967798754723,
+ "loss": 4.187763214111328,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.535672187805176,
+ "eval_runtime": 53.1736,
+ "eval_samples_per_second": 45.925,
+ "eval_steps_per_second": 1.448,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.00966129545122385,
+ "learning_rate": 0.0017722841088259045,
+ "loss": 4.1451735496521,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.010733380913734436,
+ "learning_rate": 0.0017709712304871774,
+ "loss": 4.1616363525390625,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.0117328567430377,
+ "learning_rate": 0.0017696581459170535,
+ "loss": 4.126331806182861,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.012640298344194889,
+ "learning_rate": 0.0017683448561734607,
+ "loss": 4.163730621337891,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.013074076734483242,
+ "learning_rate": 0.0017670313623144934,
+ "loss": 4.199502944946289,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.01132719125598669,
+ "learning_rate": 0.0017657176653984097,
+ "loss": 4.143380165100098,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.010008333250880241,
+ "learning_rate": 0.0017644037664836319,
+ "loss": 4.170320510864258,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.011090578511357307,
+ "learning_rate": 0.0017630896666287438,
+ "loss": 4.1571478843688965,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.011047456413507462,
+ "learning_rate": 0.0017617753668924924,
+ "loss": 4.101892948150635,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.010098653845489025,
+ "learning_rate": 0.0017604608683337857,
+ "loss": 4.154077529907227,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.010499860160052776,
+ "learning_rate": 0.00175914617201169,
+ "loss": 4.116868019104004,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.010939318686723709,
+ "learning_rate": 0.0017578312789854338,
+ "loss": 4.159144878387451,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.010622999630868435,
+ "learning_rate": 0.0017565161903144023,
+ "loss": 4.171411991119385,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.0108700692653656,
+ "learning_rate": 0.0017552009070581386,
+ "loss": 4.100748062133789,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.009794988669455051,
+ "learning_rate": 0.0017538854302763423,
+ "loss": 4.185161590576172,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.007750354241579771,
+ "learning_rate": 0.0017525697610288702,
+ "loss": 4.149083137512207,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.009923134930431843,
+ "learning_rate": 0.001751253900375732,
+ "loss": 4.205166816711426,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.010811141692101955,
+ "learning_rate": 0.0017499378493770936,
+ "loss": 4.131865978240967,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.010868688113987446,
+ "learning_rate": 0.0017486216090932733,
+ "loss": 4.148791790008545,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.011224126443266869,
+ "learning_rate": 0.001747305180584743,
+ "loss": 4.152503490447998,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.013262554071843624,
+ "learning_rate": 0.001745988564912124,
+ "loss": 4.164926528930664,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.013089599087834358,
+ "learning_rate": 0.001744671763136191,
+ "loss": 4.1626458168029785,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.01327646616846323,
+ "learning_rate": 0.0017433547763178665,
+ "loss": 4.189356803894043,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.014539042487740517,
+ "learning_rate": 0.0017420376055182233,
+ "loss": 4.176387786865234,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.013269098475575447,
+ "learning_rate": 0.001740720251798482,
+ "loss": 4.180716514587402,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.010922391898930073,
+ "learning_rate": 0.0017394027162200105,
+ "loss": 4.163928031921387,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.013355271890759468,
+ "learning_rate": 0.0017380849998443243,
+ "loss": 4.150012016296387,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.014149201102554798,
+ "learning_rate": 0.0017367671037330828,
+ "loss": 4.173713207244873,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01354879979044199,
+ "learning_rate": 0.0017354490289480918,
+ "loss": 4.150563716888428,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.013788704760372639,
+ "learning_rate": 0.0017341307765512998,
+ "loss": 4.213858604431152,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.011597222648561,
+ "learning_rate": 0.0017328123476047994,
+ "loss": 4.166989326477051,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.01055675558745861,
+ "learning_rate": 0.001731493743170824,
+ "loss": 4.172226428985596,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.011600397527217865,
+ "learning_rate": 0.0017301749643117506,
+ "loss": 4.1656951904296875,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.01114651933312416,
+ "learning_rate": 0.0017288560120900943,
+ "loss": 4.189825057983398,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.009811583906412125,
+ "learning_rate": 0.0017275368875685126,
+ "loss": 4.181164741516113,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.009758968837559223,
+ "learning_rate": 0.0017262175918097992,
+ "loss": 4.163188934326172,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.00999707542359829,
+ "learning_rate": 0.001724898125876887,
+ "loss": 4.184117317199707,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.009528180584311485,
+ "learning_rate": 0.001723578490832846,
+ "loss": 4.16407585144043,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.009717358276247978,
+ "learning_rate": 0.0017222586877408826,
+ "loss": 4.209892272949219,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.010909050703048706,
+ "learning_rate": 0.0017209387176643377,
+ "loss": 4.19977331161499,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.011234750039875507,
+ "learning_rate": 0.0017196185816666875,
+ "loss": 4.19363260269165,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.010300721041858196,
+ "learning_rate": 0.0017182982808115425,
+ "loss": 4.142427921295166,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.00941646471619606,
+ "learning_rate": 0.001716977816162644,
+ "loss": 4.159372329711914,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.009304054081439972,
+ "learning_rate": 0.0017156571887838679,
+ "loss": 4.207922458648682,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.009496180340647697,
+ "learning_rate": 0.0017143363997392184,
+ "loss": 4.137894630432129,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.009257436729967594,
+ "learning_rate": 0.0017130154500928326,
+ "loss": 4.198753833770752,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.00900586973875761,
+ "learning_rate": 0.0017116943409089746,
+ "loss": 4.160684585571289,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.009398494847118855,
+ "learning_rate": 0.00171037307325204,
+ "loss": 4.1819353103637695,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.008633013814687729,
+ "learning_rate": 0.0017090516481865483,
+ "loss": 4.176833152770996,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.00898425467312336,
+ "learning_rate": 0.0017077300667771484,
+ "loss": 4.190657138824463,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.008182398043572903,
+ "learning_rate": 0.0017064083300886155,
+ "loss": 4.173429489135742,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.009589143097400665,
+ "learning_rate": 0.0017050864391858478,
+ "loss": 4.1559014320373535,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.010304383002221584,
+ "learning_rate": 0.0017037643951338698,
+ "loss": 4.155453681945801,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.009943748824298382,
+ "learning_rate": 0.0017024421989978275,
+ "loss": 4.119734287261963,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.009238033555448055,
+ "learning_rate": 0.001701119851842992,
+ "loss": 4.188974857330322,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.008058690465986729,
+ "learning_rate": 0.0016997973547347531,
+ "loss": 4.19934606552124,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.008485953323543072,
+ "learning_rate": 0.0016984747087386233,
+ "loss": 4.165615558624268,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.008476605638861656,
+ "learning_rate": 0.0016971519149202347,
+ "loss": 4.21021842956543,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.008588708005845547,
+ "learning_rate": 0.0016958289743453385,
+ "loss": 4.208064556121826,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.009223178029060364,
+ "learning_rate": 0.0016945058880798037,
+ "loss": 4.158413887023926,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.009196167811751366,
+ "learning_rate": 0.0016931826571896166,
+ "loss": 4.203187465667725,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.009488921612501144,
+ "learning_rate": 0.001691859282740882,
+ "loss": 4.178227424621582,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.009253161028027534,
+ "learning_rate": 0.0016905357657998167,
+ "loss": 4.131230354309082,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.010069048963487148,
+ "learning_rate": 0.0016892121074327559,
+ "loss": 4.210734844207764,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.009945095516741276,
+ "learning_rate": 0.001687888308706147,
+ "loss": 4.184042930603027,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.010483640246093273,
+ "learning_rate": 0.0016865643706865505,
+ "loss": 4.156919479370117,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.011560715734958649,
+ "learning_rate": 0.0016852402944406391,
+ "loss": 4.185520172119141,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.010001277551054955,
+ "learning_rate": 0.0016839160810351978,
+ "loss": 4.226828575134277,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.010335934348404408,
+ "learning_rate": 0.0016825917315371202,
+ "loss": 4.183531284332275,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.011322776786983013,
+ "learning_rate": 0.001681267247013412,
+ "loss": 4.179635047912598,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.009663194417953491,
+ "learning_rate": 0.001679942628531186,
+ "loss": 4.165654182434082,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.010590246878564358,
+ "learning_rate": 0.0016786178771576643,
+ "loss": 4.1539459228515625,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.511386394500732,
+ "eval_runtime": 51.7751,
+ "eval_samples_per_second": 47.166,
+ "eval_steps_per_second": 1.487,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.01254354789853096,
+ "learning_rate": 0.0016772929939601728,
+ "loss": 4.104002475738525,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.01300987508147955,
+ "learning_rate": 0.0016759679800061483,
+ "loss": 4.127395153045654,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.013325816951692104,
+ "learning_rate": 0.0016746428363631299,
+ "loss": 4.131229400634766,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.014302870258688927,
+ "learning_rate": 0.001673317564098761,
+ "loss": 4.079358100891113,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.013776068575680256,
+ "learning_rate": 0.0016719921642807905,
+ "loss": 4.132065773010254,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.010788560844957829,
+ "learning_rate": 0.0016706666379770689,
+ "loss": 4.134626388549805,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.009766630828380585,
+ "learning_rate": 0.0016693409862555493,
+ "loss": 4.1645426750183105,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.011760029010474682,
+ "learning_rate": 0.0016680152101842842,
+ "loss": 4.110664367675781,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.011047836393117905,
+ "learning_rate": 0.0016666893108314288,
+ "loss": 4.129222869873047,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.009989139623939991,
+ "learning_rate": 0.0016653632892652344,
+ "loss": 4.076835632324219,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.010033783502876759,
+ "learning_rate": 0.001664037146554054,
+ "loss": 4.11555290222168,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.00996427796781063,
+ "learning_rate": 0.001662710883766337,
+ "loss": 4.131463527679443,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.01044023409485817,
+ "learning_rate": 0.0016613845019706288,
+ "loss": 4.098750114440918,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.010033872909843922,
+ "learning_rate": 0.0016600580022355712,
+ "loss": 4.161447048187256,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.009329567663371563,
+ "learning_rate": 0.0016587313856299012,
+ "loss": 4.146491527557373,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.010061589069664478,
+ "learning_rate": 0.0016574046532224504,
+ "loss": 4.107517719268799,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.009491910226643085,
+ "learning_rate": 0.0016560778060821424,
+ "loss": 4.098226547241211,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.008529345504939556,
+ "learning_rate": 0.0016547508452779943,
+ "loss": 4.1247711181640625,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.00936342403292656,
+ "learning_rate": 0.0016534237718791136,
+ "loss": 4.120573043823242,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.010255641303956509,
+ "learning_rate": 0.0016520965869547008,
+ "loss": 4.089067459106445,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.011638758704066277,
+ "learning_rate": 0.001650769291574044,
+ "loss": 4.074847221374512,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.014707857742905617,
+ "learning_rate": 0.0016494418868065214,
+ "loss": 4.153377532958984,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.01650111749768257,
+ "learning_rate": 0.0016481143737215986,
+ "loss": 4.134062767028809,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.014833372086286545,
+ "learning_rate": 0.00164678675338883,
+ "loss": 4.1344828605651855,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.012257886119186878,
+ "learning_rate": 0.0016454590268778538,
+ "loss": 4.098813533782959,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.013508322648704052,
+ "learning_rate": 0.0016441311952583959,
+ "loss": 4.106297492980957,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.01256900280714035,
+ "learning_rate": 0.001642803259600267,
+ "loss": 4.143631935119629,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.009885406121611595,
+ "learning_rate": 0.00164147522097336,
+ "loss": 4.120885848999023,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.012004674412310123,
+ "learning_rate": 0.0016401470804476518,
+ "loss": 4.089243412017822,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01224028691649437,
+ "learning_rate": 0.0016388188390932012,
+ "loss": 4.077765464782715,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.012511290609836578,
+ "learning_rate": 0.001637490497980149,
+ "loss": 4.161346435546875,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.010701716877520084,
+ "learning_rate": 0.0016361620581787143,
+ "loss": 4.10698938369751,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.010713113471865654,
+ "learning_rate": 0.001634833520759198,
+ "loss": 4.114283561706543,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.011018907651305199,
+ "learning_rate": 0.001633504886791978,
+ "loss": 4.134284019470215,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.011279584839940071,
+ "learning_rate": 0.0016321761573475117,
+ "loss": 4.1456804275512695,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.010642931796610355,
+ "learning_rate": 0.001630847333496332,
+ "loss": 4.160918235778809,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.011069678701460361,
+ "learning_rate": 0.0016295184163090482,
+ "loss": 4.136838912963867,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.011007166467607021,
+ "learning_rate": 0.0016281894068563443,
+ "loss": 4.129153728485107,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.010055703110992908,
+ "learning_rate": 0.0016268603062089802,
+ "loss": 4.138665676116943,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.009839230217039585,
+ "learning_rate": 0.0016255311154377875,
+ "loss": 4.1366496086120605,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.011191432364284992,
+ "learning_rate": 0.0016242018356136714,
+ "loss": 4.129880905151367,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.012130341492593288,
+ "learning_rate": 0.0016228724678076095,
+ "loss": 4.1425981521606445,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.012707795947790146,
+ "learning_rate": 0.0016215430130906482,
+ "loss": 4.115452766418457,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.012707809917628765,
+ "learning_rate": 0.001620213472533906,
+ "loss": 4.122758865356445,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.01182397548109293,
+ "learning_rate": 0.001618883847208569,
+ "loss": 4.140275955200195,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.010865340940654278,
+ "learning_rate": 0.001617554138185893,
+ "loss": 4.137759208679199,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.010973592288792133,
+ "learning_rate": 0.0016162243465371998,
+ "loss": 4.138947010040283,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.011247226968407631,
+ "learning_rate": 0.001614894473333879,
+ "loss": 4.115516662597656,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.010837487876415253,
+ "learning_rate": 0.0016135645196473853,
+ "loss": 4.160427093505859,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.009096396155655384,
+ "learning_rate": 0.0016122344865492385,
+ "loss": 4.1200032234191895,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.010671247728168964,
+ "learning_rate": 0.0016109043751110223,
+ "loss": 4.097919464111328,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.0107562355697155,
+ "learning_rate": 0.0016095741864043831,
+ "loss": 4.096290588378906,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.008909327909350395,
+ "learning_rate": 0.0016082439215010305,
+ "loss": 4.171017646789551,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.010305124334990978,
+ "learning_rate": 0.0016069135814727343,
+ "loss": 4.1450276374816895,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.011235613375902176,
+ "learning_rate": 0.0016055831673913255,
+ "loss": 4.134704113006592,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.00825632456690073,
+ "learning_rate": 0.0016042526803286949,
+ "loss": 4.107443332672119,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.009056378155946732,
+ "learning_rate": 0.0016029221213567921,
+ "loss": 4.132656097412109,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.008835164830088615,
+ "learning_rate": 0.001601591491547624,
+ "loss": 4.124729633331299,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.00915333442389965,
+ "learning_rate": 0.0016002607919732557,
+ "loss": 4.119500160217285,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.008532539941370487,
+ "learning_rate": 0.0015989300237058066,
+ "loss": 4.141661643981934,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.009952311404049397,
+ "learning_rate": 0.0015975991878174533,
+ "loss": 4.140929222106934,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.008920297026634216,
+ "learning_rate": 0.0015962682853804268,
+ "loss": 4.155877113342285,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.008770049549639225,
+ "learning_rate": 0.0015949373174670104,
+ "loss": 4.126394271850586,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.0078084105625748634,
+ "learning_rate": 0.0015936062851495411,
+ "loss": 4.128876209259033,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.00816641841083765,
+ "learning_rate": 0.0015922751895004078,
+ "loss": 4.137410640716553,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.008895459584891796,
+ "learning_rate": 0.0015909440315920505,
+ "loss": 4.07957649230957,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.008551528677344322,
+ "learning_rate": 0.001589612812496958,
+ "loss": 4.104213714599609,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.008325488306581974,
+ "learning_rate": 0.0015882815332876707,
+ "loss": 4.1584792137146,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.009341313503682613,
+ "learning_rate": 0.0015869501950367744,
+ "loss": 4.1417436599731445,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.00999050959944725,
+ "learning_rate": 0.0015856187988169067,
+ "loss": 4.136049270629883,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.008889932185411453,
+ "learning_rate": 0.0015842873457007484,
+ "loss": 4.16257381439209,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.009109573438763618,
+ "learning_rate": 0.0015829558367610268,
+ "loss": 4.171441555023193,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.485942840576172,
+ "eval_runtime": 52.1051,
+ "eval_samples_per_second": 46.867,
+ "eval_steps_per_second": 1.478,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.009881424717605114,
+ "learning_rate": 0.001581624273070515,
+ "loss": 4.059608459472656,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.011156721040606499,
+ "learning_rate": 0.0015802926557020294,
+ "loss": 4.047014236450195,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.012326992116868496,
+ "learning_rate": 0.0015789609857284309,
+ "loss": 4.06541109085083,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.01332208700478077,
+ "learning_rate": 0.0015776292642226213,
+ "loss": 4.042714595794678,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.012720128521323204,
+ "learning_rate": 0.0015762974922575446,
+ "loss": 4.098389625549316,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.010723215527832508,
+ "learning_rate": 0.0015749656709061854,
+ "loss": 4.059185981750488,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.010605281218886375,
+ "learning_rate": 0.001573633801241569,
+ "loss": 4.099538326263428,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.01101530622690916,
+ "learning_rate": 0.0015723018843367576,
+ "loss": 4.095674514770508,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.01135838869959116,
+ "learning_rate": 0.0015709699212648532,
+ "loss": 4.070535659790039,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.012119271792471409,
+ "learning_rate": 0.0015696379130989936,
+ "loss": 4.060605049133301,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.012783961370587349,
+ "learning_rate": 0.001568305860912355,
+ "loss": 4.075931072235107,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.011486220173537731,
+ "learning_rate": 0.0015669737657781461,
+ "loss": 4.075855255126953,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.01037552859634161,
+ "learning_rate": 0.0015656416287696136,
+ "loss": 4.087729454040527,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.009802743792533875,
+ "learning_rate": 0.0015643094509600354,
+ "loss": 4.0855607986450195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.010804438032209873,
+ "learning_rate": 0.001562977233422723,
+ "loss": 4.064528465270996,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.009505728259682655,
+ "learning_rate": 0.0015616449772310206,
+ "loss": 4.036775588989258,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.010634485632181168,
+ "learning_rate": 0.0015603126834583024,
+ "loss": 4.069923400878906,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.011392096057534218,
+ "learning_rate": 0.0015589803531779736,
+ "loss": 4.0812177658081055,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.011851722374558449,
+ "learning_rate": 0.0015576479874634683,
+ "loss": 4.09183406829834,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.011205444112420082,
+ "learning_rate": 0.0015563155873882502,
+ "loss": 4.07825231552124,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.009676349349319935,
+ "learning_rate": 0.0015549831540258093,
+ "loss": 4.052826404571533,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.009376165457069874,
+ "learning_rate": 0.0015536506884496638,
+ "loss": 4.02100133895874,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.00936955213546753,
+ "learning_rate": 0.0015523181917333564,
+ "loss": 4.049526214599609,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.00937784556299448,
+ "learning_rate": 0.0015509856649504564,
+ "loss": 4.096356391906738,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.010014008730649948,
+ "learning_rate": 0.0015496531091745555,
+ "loss": 4.03254508972168,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.009909634478390217,
+ "learning_rate": 0.0015483205254792705,
+ "loss": 4.067630767822266,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.009319501928985119,
+ "learning_rate": 0.0015469879149382399,
+ "loss": 4.094167709350586,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.008634721860289574,
+ "learning_rate": 0.0015456552786251244,
+ "loss": 4.122404098510742,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.009412454441189766,
+ "learning_rate": 0.0015443226176136049,
+ "loss": 4.081546783447266,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.010102108120918274,
+ "learning_rate": 0.0015429899329773814,
+ "loss": 4.049839019775391,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.010921654291450977,
+ "learning_rate": 0.0015416572257901746,
+ "loss": 4.097097396850586,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.010607713833451271,
+ "learning_rate": 0.0015403244971257226,
+ "loss": 4.084481239318848,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.011887960135936737,
+ "learning_rate": 0.0015389917480577806,
+ "loss": 4.074968338012695,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01089888159185648,
+ "learning_rate": 0.0015376589796601203,
+ "loss": 4.062132835388184,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.009934183210134506,
+ "learning_rate": 0.0015363261930065298,
+ "loss": 4.128092288970947,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.00878534372895956,
+ "learning_rate": 0.0015349933891708105,
+ "loss": 4.069051742553711,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.009928155690431595,
+ "learning_rate": 0.0015336605692267788,
+ "loss": 4.073912620544434,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.011091411113739014,
+ "learning_rate": 0.0015323277342482637,
+ "loss": 4.125944137573242,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.010314172133803368,
+ "learning_rate": 0.0015309948853091055,
+ "loss": 4.062620162963867,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.010423967614769936,
+ "learning_rate": 0.001529662023483158,
+ "loss": 4.101737976074219,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.009822997264564037,
+ "learning_rate": 0.0015283291498442817,
+ "loss": 4.064520359039307,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.009757095016539097,
+ "learning_rate": 0.001526996265466351,
+ "loss": 4.100624084472656,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.008650688454508781,
+ "learning_rate": 0.001525663371423246,
+ "loss": 4.109685897827148,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.007378234062343836,
+ "learning_rate": 0.0015243304687888553,
+ "loss": 4.044726848602295,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.008194336667656898,
+ "learning_rate": 0.001522997558637074,
+ "loss": 4.097923278808594,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.009764892049133778,
+ "learning_rate": 0.0015216646420418048,
+ "loss": 4.102148056030273,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.009242868050932884,
+ "learning_rate": 0.001520331720076953,
+ "loss": 4.068563461303711,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.010409324429929256,
+ "learning_rate": 0.001518998793816431,
+ "loss": 4.065944194793701,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.009215258061885834,
+ "learning_rate": 0.0015176658643341527,
+ "loss": 4.10223388671875,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.008065393194556236,
+ "learning_rate": 0.0015163329327040366,
+ "loss": 4.099432945251465,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.008498288691043854,
+ "learning_rate": 0.001515,
+ "loss": 4.109854698181152,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.008919150568544865,
+ "learning_rate": 0.0015136670672959633,
+ "loss": 4.128247261047363,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.009104429744184017,
+ "learning_rate": 0.0015123341356658474,
+ "loss": 4.068622589111328,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.008558365516364574,
+ "learning_rate": 0.0015110012061835692,
+ "loss": 4.114214897155762,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.007865076884627342,
+ "learning_rate": 0.0015096682799230477,
+ "loss": 4.11155366897583,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.007848534733057022,
+ "learning_rate": 0.0015083353579581958,
+ "loss": 4.045694351196289,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.008238030597567558,
+ "learning_rate": 0.0015070024413629264,
+ "loss": 4.093930721282959,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.007536398712545633,
+ "learning_rate": 0.0015056695312111448,
+ "loss": 4.106235504150391,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.008156892843544483,
+ "learning_rate": 0.001504336628576754,
+ "loss": 4.13060998916626,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.008144439198076725,
+ "learning_rate": 0.0015030037345336487,
+ "loss": 4.091304779052734,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.007757358253002167,
+ "learning_rate": 0.001501670850155718,
+ "loss": 4.069276809692383,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.00860623549669981,
+ "learning_rate": 0.0015003379765168425,
+ "loss": 4.107658386230469,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.008614457212388515,
+ "learning_rate": 0.0014990051146908949,
+ "loss": 4.106705665588379,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.009120592847466469,
+ "learning_rate": 0.001497672265751737,
+ "loss": 4.0891923904418945,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.009446980431675911,
+ "learning_rate": 0.0014963394307732213,
+ "loss": 4.098847389221191,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.010494766756892204,
+ "learning_rate": 0.0014950066108291894,
+ "loss": 4.115325450897217,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.010057658888399601,
+ "learning_rate": 0.0014936738069934706,
+ "loss": 4.099984169006348,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.009071771055459976,
+ "learning_rate": 0.0014923410203398796,
+ "loss": 4.091543674468994,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.009937075898051262,
+ "learning_rate": 0.00149100825194222,
+ "loss": 4.086492538452148,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.010222986340522766,
+ "learning_rate": 0.0014896755028742778,
+ "loss": 4.063841819763184,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.010513762012124062,
+ "learning_rate": 0.001488342774209826,
+ "loss": 4.11991548538208,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.009530222974717617,
+ "learning_rate": 0.0014870100670226192,
+ "loss": 4.043334007263184,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.467459678649902,
+ "eval_runtime": 54.1628,
+ "eval_samples_per_second": 45.086,
+ "eval_steps_per_second": 1.422,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.010274187661707401,
+ "learning_rate": 0.0014856773823863955,
+ "loss": 4.049561500549316,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.011448442004621029,
+ "learning_rate": 0.001484344721374876,
+ "loss": 4.019484996795654,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.011361192911863327,
+ "learning_rate": 0.0014830120850617599,
+ "loss": 4.0429887771606445,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.010250156745314598,
+ "learning_rate": 0.0014816794745207299,
+ "loss": 4.012682914733887,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.010024400427937508,
+ "learning_rate": 0.0014803468908254449,
+ "loss": 3.979409694671631,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.01153535582125187,
+ "learning_rate": 0.0014790143350495446,
+ "loss": 4.045780181884766,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.012632817961275578,
+ "learning_rate": 0.001477681808266644,
+ "loss": 4.057641983032227,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.013758293353021145,
+ "learning_rate": 0.0014763493115503366,
+ "loss": 4.061607360839844,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.014041583053767681,
+ "learning_rate": 0.001475016845974191,
+ "loss": 4.035029411315918,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.013629456982016563,
+ "learning_rate": 0.0014736844126117497,
+ "loss": 4.062169075012207,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.011023213155567646,
+ "learning_rate": 0.0014723520125365323,
+ "loss": 4.02731990814209,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.01001750584691763,
+ "learning_rate": 0.0014710196468220266,
+ "loss": 4.019882678985596,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.008321646600961685,
+ "learning_rate": 0.001469687316541698,
+ "loss": 4.005446434020996,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.008935946971178055,
+ "learning_rate": 0.0014683550227689796,
+ "loss": 4.017102241516113,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.009137130342423916,
+ "learning_rate": 0.001467022766577277,
+ "loss": 4.024102210998535,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.00805022194981575,
+ "learning_rate": 0.001465690549039965,
+ "loss": 3.9935250282287598,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.009204058907926083,
+ "learning_rate": 0.0014643583712303865,
+ "loss": 4.042818546295166,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.009812815114855766,
+ "learning_rate": 0.001463026234221854,
+ "loss": 4.0252485275268555,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.009955782443284988,
+ "learning_rate": 0.0014616941390876454,
+ "loss": 3.9839768409729004,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.009374645538628101,
+ "learning_rate": 0.001460362086901007,
+ "loss": 4.040245056152344,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.010008739307522774,
+ "learning_rate": 0.0014590300787351471,
+ "loss": 4.0306596755981445,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.009014200419187546,
+ "learning_rate": 0.001457698115663243,
+ "loss": 4.028713226318359,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.008381523191928864,
+ "learning_rate": 0.0014563661987584313,
+ "loss": 4.024718284606934,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.008627948351204395,
+ "learning_rate": 0.0014550343290938143,
+ "loss": 4.02303409576416,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.010257705114781857,
+ "learning_rate": 0.0014537025077424556,
+ "loss": 4.061058044433594,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.010759209282696247,
+ "learning_rate": 0.0014523707357773788,
+ "loss": 4.064718246459961,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.009966139681637287,
+ "learning_rate": 0.0014510390142715697,
+ "loss": 3.9919447898864746,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.00997584406286478,
+ "learning_rate": 0.001449707344297971,
+ "loss": 4.0431599617004395,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.010846944525837898,
+ "learning_rate": 0.0014483757269294858,
+ "loss": 4.0222697257995605,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.00991352554410696,
+ "learning_rate": 0.0014470441632389735,
+ "loss": 4.059438705444336,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.008994397707283497,
+ "learning_rate": 0.0014457126542992518,
+ "loss": 4.021494388580322,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.009567267261445522,
+ "learning_rate": 0.0014443812011830934,
+ "loss": 4.070810317993164,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.009199011139571667,
+ "learning_rate": 0.0014430498049632255,
+ "loss": 4.020938873291016,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.010440350510179996,
+ "learning_rate": 0.0014417184667123303,
+ "loss": 4.062529563903809,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.009962482377886772,
+ "learning_rate": 0.0014403871875030424,
+ "loss": 4.0687150955200195,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.009389379993081093,
+ "learning_rate": 0.0014390559684079502,
+ "loss": 4.055185794830322,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.011469941586256027,
+ "learning_rate": 0.0014377248104995924,
+ "loss": 4.041023254394531,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.011034650728106499,
+ "learning_rate": 0.0014363937148504588,
+ "loss": 4.054254531860352,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.010651151649653912,
+ "learning_rate": 0.0014350626825329897,
+ "loss": 4.037753582000732,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.010747930034995079,
+ "learning_rate": 0.0014337317146195733,
+ "loss": 4.059938907623291,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.00987576600164175,
+ "learning_rate": 0.0014324008121825468,
+ "loss": 4.075366973876953,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.008504371158778667,
+ "learning_rate": 0.0014310699762941936,
+ "loss": 4.026755332946777,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.009445052593946457,
+ "learning_rate": 0.001429739208026745,
+ "loss": 4.034431457519531,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.010560023598372936,
+ "learning_rate": 0.0014284085084523763,
+ "loss": 4.101158618927002,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.009920580312609673,
+ "learning_rate": 0.0014270778786432078,
+ "loss": 4.041321754455566,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.009854268282651901,
+ "learning_rate": 0.0014257473196713053,
+ "loss": 4.049882411956787,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.009825135581195354,
+ "learning_rate": 0.0014244168326086745,
+ "loss": 4.067440986633301,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.010506784543395042,
+ "learning_rate": 0.0014230864185272663,
+ "loss": 4.042476654052734,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.009527767077088356,
+ "learning_rate": 0.0014217560784989696,
+ "loss": 4.055248260498047,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.01013826485723257,
+ "learning_rate": 0.001420425813595617,
+ "loss": 4.05246639251709,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.011020111851394176,
+ "learning_rate": 0.0014190956248889778,
+ "loss": 4.032987117767334,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.010647117160260677,
+ "learning_rate": 0.0014177655134507614,
+ "loss": 4.093589782714844,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.009484780952334404,
+ "learning_rate": 0.001416435480352615,
+ "loss": 4.017437934875488,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.0114199910312891,
+ "learning_rate": 0.0014151055266661211,
+ "loss": 4.080643653869629,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.011137664318084717,
+ "learning_rate": 0.0014137756534628007,
+ "loss": 4.069380760192871,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.008972682990133762,
+ "learning_rate": 0.0014124458618141073,
+ "loss": 4.051383018493652,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.010311643593013287,
+ "learning_rate": 0.0014111161527914317,
+ "loss": 4.048953533172607,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.01113453321158886,
+ "learning_rate": 0.0014097865274660944,
+ "loss": 4.063544273376465,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.01091703586280346,
+ "learning_rate": 0.0014084569869093518,
+ "loss": 4.006721496582031,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.010672449134290218,
+ "learning_rate": 0.0014071275321923906,
+ "loss": 4.020220756530762,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.009682527743279934,
+ "learning_rate": 0.001405798164386328,
+ "loss": 4.057307243347168,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.009808676317334175,
+ "learning_rate": 0.0014044688845622126,
+ "loss": 4.105412483215332,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.010351189412176609,
+ "learning_rate": 0.0014031396937910202,
+ "loss": 4.066454887390137,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.010038275271654129,
+ "learning_rate": 0.001401810593143656,
+ "loss": 4.056752681732178,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01016094908118248,
+ "learning_rate": 0.001400481583690952,
+ "loss": 4.097933292388916,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.008683547377586365,
+ "learning_rate": 0.0013991526665036679,
+ "loss": 4.0386152267456055,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.009556332603096962,
+ "learning_rate": 0.0013978238426524882,
+ "loss": 4.05835485458374,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.009597044438123703,
+ "learning_rate": 0.001396495113208022,
+ "loss": 4.052363872528076,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.010021139867603779,
+ "learning_rate": 0.0013951664792408026,
+ "loss": 4.083431243896484,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.010611358098685741,
+ "learning_rate": 0.0013938379418212863,
+ "loss": 4.042330265045166,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.009132112376391888,
+ "learning_rate": 0.001392509502019852,
+ "loss": 4.058825969696045,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.009808709844946861,
+ "learning_rate": 0.0013911811609067992,
+ "loss": 4.093010902404785,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.4473161697387695,
+ "eval_runtime": 51.7738,
+ "eval_samples_per_second": 47.167,
+ "eval_steps_per_second": 1.487,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.011859052814543247,
+ "learning_rate": 0.0013898529195523488,
+ "loss": 3.9572651386260986,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.013210238888859749,
+ "learning_rate": 0.0013885247790266406,
+ "loss": 3.9812350273132324,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.012981324456632137,
+ "learning_rate": 0.001387196740399733,
+ "loss": 3.987725019454956,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.01424382347613573,
+ "learning_rate": 0.001385868804741604,
+ "loss": 3.9719889163970947,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.0154642965644598,
+ "learning_rate": 0.0013845409731221465,
+ "loss": 3.984248161315918,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.014779271557927132,
+ "learning_rate": 0.0013832132466111706,
+ "loss": 4.002912521362305,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.012921188026666641,
+ "learning_rate": 0.0013818856262784013,
+ "loss": 3.9588499069213867,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01165325753390789,
+ "learning_rate": 0.0013805581131934792,
+ "loss": 3.993288993835449,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.008908954448997974,
+ "learning_rate": 0.0013792307084259563,
+ "loss": 3.987151861190796,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.012040507979691029,
+ "learning_rate": 0.0013779034130452991,
+ "loss": 3.9675703048706055,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.013464349322021008,
+ "learning_rate": 0.0013765762281208868,
+ "loss": 3.9901437759399414,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.009626250714063644,
+ "learning_rate": 0.0013752491547220063,
+ "loss": 3.9373281002044678,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.00897257961332798,
+ "learning_rate": 0.0013739221939178582,
+ "loss": 4.00804328918457,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.010689414106309414,
+ "learning_rate": 0.00137259534677755,
+ "loss": 4.020804405212402,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.010462156496942043,
+ "learning_rate": 0.0013712686143700988,
+ "loss": 4.028973579406738,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.009135857224464417,
+ "learning_rate": 0.001369941997764429,
+ "loss": 3.9798216819763184,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.009298899210989475,
+ "learning_rate": 0.0013686154980293713,
+ "loss": 4.018845081329346,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.00979928020387888,
+ "learning_rate": 0.0013672891162336636,
+ "loss": 3.9946775436401367,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.008175542578101158,
+ "learning_rate": 0.0013659628534459457,
+ "loss": 3.981858491897583,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.008302642032504082,
+ "learning_rate": 0.0013646367107347662,
+ "loss": 3.991128444671631,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.008516009896993637,
+ "learning_rate": 0.0013633106891685718,
+ "loss": 4.004990100860596,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.007936482317745686,
+ "learning_rate": 0.0013619847898157164,
+ "loss": 3.995734214782715,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.008317600004374981,
+ "learning_rate": 0.001360659013744451,
+ "loss": 4.040653228759766,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.008481282740831375,
+ "learning_rate": 0.001359333362022931,
+ "loss": 3.969860553741455,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.008944948203861713,
+ "learning_rate": 0.0013580078357192096,
+ "loss": 3.9908909797668457,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.008897850289940834,
+ "learning_rate": 0.001356682435901239,
+ "loss": 4.04786491394043,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.009527456946671009,
+ "learning_rate": 0.001355357163636871,
+ "loss": 4.017423629760742,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.011714582331478596,
+ "learning_rate": 0.001354032019993852,
+ "loss": 3.9990692138671875,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.011083807796239853,
+ "learning_rate": 0.0013527070060398274,
+ "loss": 3.995265483856201,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.009878288954496384,
+ "learning_rate": 0.0013513821228423363,
+ "loss": 4.0364532470703125,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.009727329947054386,
+ "learning_rate": 0.0013500573714688136,
+ "loss": 4.035953521728516,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.009482033550739288,
+ "learning_rate": 0.0013487327529865879,
+ "loss": 4.015855312347412,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.009037062525749207,
+ "learning_rate": 0.0013474082684628798,
+ "loss": 4.024164199829102,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.009094092063605785,
+ "learning_rate": 0.0013460839189648028,
+ "loss": 4.037765979766846,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.008230654522776604,
+ "learning_rate": 0.0013447597055593612,
+ "loss": 4.006317615509033,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.008491289801895618,
+ "learning_rate": 0.0013434356293134501,
+ "loss": 4.042603969573975,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.008907693438231945,
+ "learning_rate": 0.0013421116912938533,
+ "loss": 4.008098602294922,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.008897818624973297,
+ "learning_rate": 0.0013407878925672438,
+ "loss": 3.9892945289611816,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.00957555789500475,
+ "learning_rate": 0.0013394642342001832,
+ "loss": 4.046009063720703,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.009948568418622017,
+ "learning_rate": 0.0013381407172591185,
+ "loss": 3.9760642051696777,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.01014955248683691,
+ "learning_rate": 0.0013368173428103835,
+ "loss": 4.023215293884277,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.008578895591199398,
+ "learning_rate": 0.0013354941119201966,
+ "loss": 4.002819061279297,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.008729875087738037,
+ "learning_rate": 0.001334171025654662,
+ "loss": 4.0340657234191895,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.009556280449032784,
+ "learning_rate": 0.0013328480850797655,
+ "loss": 3.983076333999634,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.009733820334076881,
+ "learning_rate": 0.0013315252912613766,
+ "loss": 4.066578388214111,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.009025252424180508,
+ "learning_rate": 0.0013302026452652474,
+ "loss": 3.9980697631835938,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.008485320024192333,
+ "learning_rate": 0.001328880148157008,
+ "loss": 4.018826961517334,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.009539696387946606,
+ "learning_rate": 0.0013275578010021726,
+ "loss": 4.011737823486328,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.010881016030907631,
+ "learning_rate": 0.0013262356048661305,
+ "loss": 4.007291793823242,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.010362236760556698,
+ "learning_rate": 0.0013249135608141523,
+ "loss": 3.9929394721984863,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.010296608321368694,
+ "learning_rate": 0.0013235916699113849,
+ "loss": 4.0015130043029785,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.011211174540221691,
+ "learning_rate": 0.0013222699332228517,
+ "loss": 4.005324363708496,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.009743994101881981,
+ "learning_rate": 0.0013209483518134525,
+ "loss": 4.0278143882751465,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.008563914336264133,
+ "learning_rate": 0.0013196269267479607,
+ "loss": 3.9997801780700684,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.009078902192413807,
+ "learning_rate": 0.0013183056590910255,
+ "loss": 4.026200294494629,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.009618895128369331,
+ "learning_rate": 0.0013169845499071675,
+ "loss": 4.027050018310547,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.008920644409954548,
+ "learning_rate": 0.0013156636002607822,
+ "loss": 4.000490665435791,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.008849202655255795,
+ "learning_rate": 0.0013143428112161323,
+ "loss": 4.044985294342041,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.009303621016442776,
+ "learning_rate": 0.001313022183837356,
+ "loss": 4.046281814575195,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.01076227892190218,
+ "learning_rate": 0.0013117017191884579,
+ "loss": 4.030576705932617,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.010400841943919659,
+ "learning_rate": 0.0013103814183333127,
+ "loss": 3.980081558227539,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.009459166787564754,
+ "learning_rate": 0.001309061282335663,
+ "loss": 4.014824867248535,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.009054969064891338,
+ "learning_rate": 0.001307741312259118,
+ "loss": 4.0485944747924805,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.009003991261124611,
+ "learning_rate": 0.0013064215091671545,
+ "loss": 3.9990320205688477,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.008959448896348476,
+ "learning_rate": 0.001305101874123113,
+ "loss": 4.038620948791504,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.009102243930101395,
+ "learning_rate": 0.0013037824081902015,
+ "loss": 4.030906677246094,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.008599337190389633,
+ "learning_rate": 0.0013024631124314877,
+ "loss": 4.02648401260376,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.008771977387368679,
+ "learning_rate": 0.0013011439879099056,
+ "loss": 4.027151107788086,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.009082595817744732,
+ "learning_rate": 0.00129982503568825,
+ "loss": 4.0194478034973145,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.009025354869663715,
+ "learning_rate": 0.0012985062568291763,
+ "loss": 4.044382095336914,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.00875102635473013,
+ "learning_rate": 0.0012971876523952014,
+ "loss": 4.0257568359375,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.009312704205513,
+ "learning_rate": 0.0012958692234487006,
+ "loss": 3.9964494705200195,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.431707859039307,
+ "eval_runtime": 50.7311,
+ "eval_samples_per_second": 48.136,
+ "eval_steps_per_second": 1.518,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.00992550328373909,
+ "learning_rate": 0.0012945509710519088,
+ "loss": 3.9253792762756348,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.009920453652739525,
+ "learning_rate": 0.0012932328962669173,
+ "loss": 3.9438819885253906,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.009437218308448792,
+ "learning_rate": 0.0012919150001556756,
+ "loss": 3.960963249206543,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.009616893716156483,
+ "learning_rate": 0.0012905972837799896,
+ "loss": 3.969301223754883,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.009107022546231747,
+ "learning_rate": 0.0012892797482015185,
+ "loss": 3.9626078605651855,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.011317167431116104,
+ "learning_rate": 0.0012879623944817772,
+ "loss": 3.946483850479126,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.010536770336329937,
+ "learning_rate": 0.001286645223682134,
+ "loss": 3.959566116333008,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.009349033236503601,
+ "learning_rate": 0.0012853282368638097,
+ "loss": 3.964435338973999,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.009950080886483192,
+ "learning_rate": 0.0012840114350878762,
+ "loss": 3.9729413986206055,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.010833023115992546,
+ "learning_rate": 0.001282694819415257,
+ "loss": 4.006421089172363,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.011538525111973286,
+ "learning_rate": 0.0012813783909067266,
+ "loss": 3.9558355808258057,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.010671315714716911,
+ "learning_rate": 0.0012800621506229065,
+ "loss": 4.0052337646484375,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.010835555382072926,
+ "learning_rate": 0.0012787460996242686,
+ "loss": 3.958641767501831,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.01082367543131113,
+ "learning_rate": 0.0012774302389711304,
+ "loss": 3.9818902015686035,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.009194346144795418,
+ "learning_rate": 0.0012761145697236583,
+ "loss": 3.9688854217529297,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.007859935984015465,
+ "learning_rate": 0.001274799092941862,
+ "loss": 3.9730639457702637,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.008534743450582027,
+ "learning_rate": 0.0012734838096855978,
+ "loss": 3.975003719329834,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.009369445964694023,
+ "learning_rate": 0.0012721687210145666,
+ "loss": 3.972907304763794,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.0090147340670228,
+ "learning_rate": 0.0012708538279883098,
+ "loss": 3.964256525039673,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.007745139300823212,
+ "learning_rate": 0.0012695391316662151,
+ "loss": 3.9504833221435547,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.008920101448893547,
+ "learning_rate": 0.0012682246331075075,
+ "loss": 3.9735097885131836,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.008987226523458958,
+ "learning_rate": 0.0012669103333712568,
+ "loss": 3.9482927322387695,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.00841289572417736,
+ "learning_rate": 0.0012655962335163685,
+ "loss": 3.965221881866455,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.008016116917133331,
+ "learning_rate": 0.0012642823346015902,
+ "loss": 3.9390206336975098,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.008470152504742146,
+ "learning_rate": 0.0012629686376855068,
+ "loss": 3.9257445335388184,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.008637234568595886,
+ "learning_rate": 0.0012616551438265394,
+ "loss": 3.9999356269836426,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.008036923594772816,
+ "learning_rate": 0.0012603418540829473,
+ "loss": 3.9755172729492188,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.008908843621611595,
+ "learning_rate": 0.0012590287695128232,
+ "loss": 3.9248545169830322,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.008743193000555038,
+ "learning_rate": 0.001257715891174096,
+ "loss": 3.9619390964508057,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.00834199320524931,
+ "learning_rate": 0.001256403220124528,
+ "loss": 3.961569309234619,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.009148893877863884,
+ "learning_rate": 0.001255090757421715,
+ "loss": 3.972104549407959,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.00925084762275219,
+ "learning_rate": 0.0012537785041230843,
+ "loss": 4.000775337219238,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.008659154176712036,
+ "learning_rate": 0.0012524664612858947,
+ "loss": 3.954564332962036,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.007787859998643398,
+ "learning_rate": 0.0012511546299672355,
+ "loss": 3.9498860836029053,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.00817918125540018,
+ "learning_rate": 0.0012498430112240246,
+ "loss": 4.003897666931152,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.009282548911869526,
+ "learning_rate": 0.001248531606113011,
+ "loss": 3.9621925354003906,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.009375610388815403,
+ "learning_rate": 0.001247220415690769,
+ "loss": 3.9977524280548096,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.009650378488004208,
+ "learning_rate": 0.001245909441013701,
+ "loss": 3.9882829189300537,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.009350919164717197,
+ "learning_rate": 0.0012445986831380366,
+ "loss": 4.019596099853516,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.008726303465664387,
+ "learning_rate": 0.0012432881431198289,
+ "loss": 3.9933621883392334,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.00829484686255455,
+ "learning_rate": 0.0012419778220149568,
+ "loss": 4.001237869262695,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.008744452148675919,
+ "learning_rate": 0.0012406677208791218,
+ "loss": 3.9323887825012207,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.008170492947101593,
+ "learning_rate": 0.0012393578407678496,
+ "loss": 3.925581455230713,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.007843422703444958,
+ "learning_rate": 0.001238048182736486,
+ "loss": 4.001023769378662,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.008914309553802013,
+ "learning_rate": 0.0012367387478401987,
+ "loss": 3.9804537296295166,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.009736751206219196,
+ "learning_rate": 0.0012354295371339774,
+ "loss": 3.9831767082214355,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.010061381384730339,
+ "learning_rate": 0.0012341205516726273,
+ "loss": 3.9462435245513916,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.010350692085921764,
+ "learning_rate": 0.0012328117925107765,
+ "loss": 4.012863636016846,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01039902213960886,
+ "learning_rate": 0.0012315032607028667,
+ "loss": 3.9811301231384277,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.011423371732234955,
+ "learning_rate": 0.0012301949573031595,
+ "loss": 3.9773542881011963,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.012449614703655243,
+ "learning_rate": 0.001228886883365731,
+ "loss": 3.969600200653076,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.010361429303884506,
+ "learning_rate": 0.0012275790399444726,
+ "loss": 4.035360336303711,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.009603082202374935,
+ "learning_rate": 0.0012262714280930904,
+ "loss": 3.9840846061706543,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.009183845482766628,
+ "learning_rate": 0.0012249640488651036,
+ "loss": 4.031542778015137,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.009694446809589863,
+ "learning_rate": 0.0012236569033138441,
+ "loss": 3.9935057163238525,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.0097816726192832,
+ "learning_rate": 0.0012223499924924545,
+ "loss": 3.9851176738739014,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.009413184598088264,
+ "learning_rate": 0.001221043317453891,
+ "loss": 3.9642205238342285,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.009960711002349854,
+ "learning_rate": 0.0012197368792509166,
+ "loss": 3.9373817443847656,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.010124670341610909,
+ "learning_rate": 0.0012184306789361059,
+ "loss": 3.958115577697754,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.011087656021118164,
+ "learning_rate": 0.00121712471756184,
+ "loss": 3.9607057571411133,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.009812800213694572,
+ "learning_rate": 0.001215818996180309,
+ "loss": 4.00722599029541,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.010305573232471943,
+ "learning_rate": 0.0012145135158435094,
+ "loss": 3.98728084564209,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.01019787136465311,
+ "learning_rate": 0.0012132082776032422,
+ "loss": 3.981172800064087,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.009260181337594986,
+ "learning_rate": 0.001211903282511115,
+ "loss": 4.034339904785156,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.00972998607903719,
+ "learning_rate": 0.001210598531618538,
+ "loss": 4.022885322570801,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.009853586554527283,
+ "learning_rate": 0.0012092940259767267,
+ "loss": 4.000235080718994,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.00961449183523655,
+ "learning_rate": 0.0012079897666366962,
+ "loss": 3.9951353073120117,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.010257176123559475,
+ "learning_rate": 0.0012066857546492657,
+ "loss": 3.9801371097564697,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.009443561546504498,
+ "learning_rate": 0.001205381991065054,
+ "loss": 3.9596502780914307,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.010389018803834915,
+ "learning_rate": 0.0012040784769344795,
+ "loss": 3.980527639389038,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.010322567075490952,
+ "learning_rate": 0.0012027752133077607,
+ "loss": 3.996258497238159,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.009987728670239449,
+ "learning_rate": 0.0012014722012349125,
+ "loss": 3.9493203163146973,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.421086311340332,
+ "eval_runtime": 52.5623,
+ "eval_samples_per_second": 46.459,
+ "eval_steps_per_second": 1.465,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.009405115619301796,
+ "learning_rate": 0.0012001694417657493,
+ "loss": 3.927506923675537,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.009974643588066101,
+ "learning_rate": 0.00119886693594988,
+ "loss": 3.9373550415039062,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.011331096291542053,
+ "learning_rate": 0.0011975646848367098,
+ "loss": 3.935594081878662,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.010538448579609394,
+ "learning_rate": 0.0011962626894754405,
+ "loss": 3.911579132080078,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.00989572424441576,
+ "learning_rate": 0.0011949609509150645,
+ "loss": 3.92525577545166,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.010533696971833706,
+ "learning_rate": 0.0011936594702043704,
+ "loss": 3.898075580596924,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.01051455456763506,
+ "learning_rate": 0.0011923582483919365,
+ "loss": 3.93463134765625,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.009662075899541378,
+ "learning_rate": 0.0011910572865261348,
+ "loss": 3.898221731185913,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.008874096907675266,
+ "learning_rate": 0.0011897565856551257,
+ "loss": 3.926787853240967,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.008126807399094105,
+ "learning_rate": 0.0011884561468268606,
+ "loss": 3.9158413410186768,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.007915553636848927,
+ "learning_rate": 0.001187155971089081,
+ "loss": 3.8891496658325195,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.008312922902405262,
+ "learning_rate": 0.0011858560594893129,
+ "loss": 3.932219982147217,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.008892328478395939,
+ "learning_rate": 0.0011845564130748734,
+ "loss": 3.8866336345672607,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.009187526069581509,
+ "learning_rate": 0.0011832570328928628,
+ "loss": 3.9464707374572754,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.008615494705736637,
+ "learning_rate": 0.0011819579199901693,
+ "loss": 3.9396629333496094,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.008649308234453201,
+ "learning_rate": 0.0011806590754134636,
+ "loss": 3.9325060844421387,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.008581082336604595,
+ "learning_rate": 0.0011793605002092021,
+ "loss": 3.9200944900512695,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.008502699434757233,
+ "learning_rate": 0.0011780621954236238,
+ "loss": 3.9656100273132324,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.0083576999604702,
+ "learning_rate": 0.0011767641621027477,
+ "loss": 3.9130215644836426,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.008196269161999226,
+ "learning_rate": 0.0011754664012923785,
+ "loss": 3.942369222640991,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.008857307955622673,
+ "learning_rate": 0.0011741689140380956,
+ "loss": 3.909585952758789,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.008935349993407726,
+ "learning_rate": 0.0011728717013852635,
+ "loss": 3.9248366355895996,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.008821200579404831,
+ "learning_rate": 0.0011715747643790211,
+ "loss": 3.9589288234710693,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.00861942209303379,
+ "learning_rate": 0.0011702781040642885,
+ "loss": 3.9381728172302246,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.009127574972808361,
+ "learning_rate": 0.0011689817214857613,
+ "loss": 3.897829532623291,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.008008269593119621,
+ "learning_rate": 0.0011676856176879107,
+ "loss": 3.930056095123291,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.010070208460092545,
+ "learning_rate": 0.0011663897937149853,
+ "loss": 3.892270803451538,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.00903113093227148,
+ "learning_rate": 0.001165094250611006,
+ "loss": 3.9766440391540527,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.008872415870428085,
+ "learning_rate": 0.0011637989894197692,
+ "loss": 3.9559712409973145,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.009524816647171974,
+ "learning_rate": 0.0011625040111848427,
+ "loss": 3.9607882499694824,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.009836779907345772,
+ "learning_rate": 0.0011612093169495678,
+ "loss": 3.9629592895507812,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.009586540050804615,
+ "learning_rate": 0.0011599149077570561,
+ "loss": 3.95208740234375,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.010495901107788086,
+ "learning_rate": 0.0011586207846501896,
+ "loss": 3.9439730644226074,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.010346266441047192,
+ "learning_rate": 0.0011573269486716202,
+ "loss": 3.959299087524414,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.010767634026706219,
+ "learning_rate": 0.0011560334008637674,
+ "loss": 3.900784969329834,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.012122200801968575,
+ "learning_rate": 0.0011547401422688202,
+ "loss": 3.9588193893432617,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.012914645485579967,
+ "learning_rate": 0.0011534471739287331,
+ "loss": 3.9278407096862793,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.011785333044826984,
+ "learning_rate": 0.0011521544968852273,
+ "loss": 3.992793560028076,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.008455781266093254,
+ "learning_rate": 0.0011508621121797899,
+ "loss": 3.96525239944458,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.009655282832682133,
+ "learning_rate": 0.0011495700208536714,
+ "loss": 3.9346890449523926,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.010537004098296165,
+ "learning_rate": 0.0011482782239478871,
+ "loss": 3.9534668922424316,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.009438651613891125,
+ "learning_rate": 0.0011469867225032135,
+ "loss": 3.983487606048584,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.010586646385490894,
+ "learning_rate": 0.0011456955175601908,
+ "loss": 3.939680576324463,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.009082718752324581,
+ "learning_rate": 0.001144404610159119,
+ "loss": 3.9558515548706055,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.009833602234721184,
+ "learning_rate": 0.0011431140013400591,
+ "loss": 3.9539828300476074,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.009142170660197735,
+ "learning_rate": 0.0011418236921428313,
+ "loss": 3.9442849159240723,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.008603853173553944,
+ "learning_rate": 0.0011405336836070137,
+ "loss": 3.9526913166046143,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.008838302455842495,
+ "learning_rate": 0.0011392439767719446,
+ "loss": 3.973104953765869,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.008867006748914719,
+ "learning_rate": 0.0011379545726767156,
+ "loss": 3.9832305908203125,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.00971697736531496,
+ "learning_rate": 0.001136665472360178,
+ "loss": 3.9563937187194824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.00957175437361002,
+ "learning_rate": 0.001135376676860935,
+ "loss": 3.975194215774536,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.008548660203814507,
+ "learning_rate": 0.001134088187217347,
+ "loss": 3.9839985370635986,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.008556334301829338,
+ "learning_rate": 0.0011328000044675264,
+ "loss": 3.9225516319274902,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.008440867997705936,
+ "learning_rate": 0.0011315121296493385,
+ "loss": 3.945028066635132,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.009643333032727242,
+ "learning_rate": 0.0011302245638004017,
+ "loss": 3.943120002746582,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.008832835592329502,
+ "learning_rate": 0.001128937307958083,
+ "loss": 3.9895920753479004,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.008720474317669868,
+ "learning_rate": 0.0011276503631595036,
+ "loss": 3.941950798034668,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.008099926635622978,
+ "learning_rate": 0.001126363730441529,
+ "loss": 3.9626870155334473,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.008525107055902481,
+ "learning_rate": 0.0011250774108407778,
+ "loss": 3.950535774230957,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.008998272009193897,
+ "learning_rate": 0.0011237914053936135,
+ "loss": 3.9460678100585938,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.009387437254190445,
+ "learning_rate": 0.001122505715136148,
+ "loss": 3.926604986190796,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.007804017513990402,
+ "learning_rate": 0.0011212203411042386,
+ "loss": 3.9894230365753174,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.008564616553485394,
+ "learning_rate": 0.0011199352843334876,
+ "loss": 3.9339959621429443,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.008726440370082855,
+ "learning_rate": 0.0011186505458592427,
+ "loss": 3.9867987632751465,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.008074100129306316,
+ "learning_rate": 0.0011173661267165932,
+ "loss": 3.9492292404174805,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.007688041776418686,
+ "learning_rate": 0.0011160820279403744,
+ "loss": 3.9425454139709473,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.007808993570506573,
+ "learning_rate": 0.0011147982505651595,
+ "loss": 3.935114860534668,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.009786820039153099,
+ "learning_rate": 0.0011135147956252659,
+ "loss": 3.959758758544922,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.009487155824899673,
+ "learning_rate": 0.00111223166415475,
+ "loss": 3.989048480987549,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.008407804183661938,
+ "learning_rate": 0.0011109488571874073,
+ "loss": 3.932347297668457,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.007158380467444658,
+ "learning_rate": 0.0011096663757567723,
+ "loss": 3.9545154571533203,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.008840581402182579,
+ "learning_rate": 0.001108384220896117,
+ "loss": 3.965984344482422,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.409194469451904,
+ "eval_runtime": 51.5561,
+ "eval_samples_per_second": 47.366,
+ "eval_steps_per_second": 1.494,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.010702896863222122,
+ "learning_rate": 0.0011071023936384508,
+ "loss": 3.8964462280273438,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01290619745850563,
+ "learning_rate": 0.0011058208950165183,
+ "loss": 3.8775625228881836,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01443667896091938,
+ "learning_rate": 0.0011045397260628,
+ "loss": 3.8755288124084473,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.014411968179047108,
+ "learning_rate": 0.0011032588878095108,
+ "loss": 3.889918327331543,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.013656266033649445,
+ "learning_rate": 0.0011019783812885986,
+ "loss": 3.926997661590576,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.010661937296390533,
+ "learning_rate": 0.0011006982075317446,
+ "loss": 3.9226956367492676,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.008968768641352654,
+ "learning_rate": 0.0010994183675703615,
+ "loss": 3.8932647705078125,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.009992181323468685,
+ "learning_rate": 0.0010981388624355937,
+ "loss": 3.8920068740844727,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.011034460738301277,
+ "learning_rate": 0.0010968596931583148,
+ "loss": 3.909073829650879,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.011815821751952171,
+ "learning_rate": 0.0010955808607691284,
+ "loss": 3.9190006256103516,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.012317490763962269,
+ "learning_rate": 0.0010943023662983674,
+ "loss": 3.8790552616119385,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.010983958840370178,
+ "learning_rate": 0.0010930242107760903,
+ "loss": 3.8730852603912354,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.008948802947998047,
+ "learning_rate": 0.0010917463952320856,
+ "loss": 3.923659563064575,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.009568693116307259,
+ "learning_rate": 0.0010904689206958649,
+ "loss": 3.9100794792175293,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.011862963438034058,
+ "learning_rate": 0.0010891917881966672,
+ "loss": 3.908093214035034,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.010180995799601078,
+ "learning_rate": 0.0010879149987634545,
+ "loss": 3.900660514831543,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.009372464381158352,
+ "learning_rate": 0.0010866385534249131,
+ "loss": 3.8950703144073486,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.010695629753172398,
+ "learning_rate": 0.0010853624532094525,
+ "loss": 3.913530111312866,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.011817459017038345,
+ "learning_rate": 0.0010840866991452024,
+ "loss": 3.90268611907959,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.010686337016522884,
+ "learning_rate": 0.0010828112922600167,
+ "loss": 3.9141364097595215,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.009550466202199459,
+ "learning_rate": 0.0010815362335814655,
+ "loss": 3.968468189239502,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.009367208927869797,
+ "learning_rate": 0.0010802615241368423,
+ "loss": 3.8933968544006348,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.009899397380650043,
+ "learning_rate": 0.0010789871649531562,
+ "loss": 3.936708450317383,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.009181005880236626,
+ "learning_rate": 0.0010777131570571363,
+ "loss": 3.9370503425598145,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.00799303874373436,
+ "learning_rate": 0.0010764395014752274,
+ "loss": 3.911050319671631,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.008469315245747566,
+ "learning_rate": 0.0010751661992335908,
+ "loss": 3.9103240966796875,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.009814459830522537,
+ "learning_rate": 0.0010738932513581034,
+ "loss": 3.881843090057373,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.010244108736515045,
+ "learning_rate": 0.0010726206588743557,
+ "loss": 3.883141040802002,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.008898732252418995,
+ "learning_rate": 0.001071348422807653,
+ "loss": 3.952347755432129,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.00823870487511158,
+ "learning_rate": 0.0010700765441830121,
+ "loss": 3.932507038116455,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.009857657365500927,
+ "learning_rate": 0.001068805024025164,
+ "loss": 3.932932138442993,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.00929308496415615,
+ "learning_rate": 0.0010675338633585488,
+ "loss": 3.8971338272094727,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.008667961694300175,
+ "learning_rate": 0.0010662630632073172,
+ "loss": 3.916931629180908,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.009758287109434605,
+ "learning_rate": 0.001064992624595331,
+ "loss": 3.964726209640503,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.009429254569113255,
+ "learning_rate": 0.0010637225485461586,
+ "loss": 3.917572498321533,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.008303353562951088,
+ "learning_rate": 0.0010624528360830782,
+ "loss": 3.875241756439209,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.008729013614356518,
+ "learning_rate": 0.001061183488229073,
+ "loss": 3.9691600799560547,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.010239302180707455,
+ "learning_rate": 0.0010599145060068353,
+ "loss": 3.9053149223327637,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.00932721421122551,
+ "learning_rate": 0.0010586458904387596,
+ "loss": 3.8593955039978027,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.008196988143026829,
+ "learning_rate": 0.0010573776425469473,
+ "loss": 3.917983055114746,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.008049141615629196,
+ "learning_rate": 0.001056109763353203,
+ "loss": 3.9337785243988037,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.009256313554942608,
+ "learning_rate": 0.001054842253879034,
+ "loss": 3.937887668609619,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.009292179718613625,
+ "learning_rate": 0.0010535751151456495,
+ "loss": 3.9203450679779053,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.009142161346971989,
+ "learning_rate": 0.0010523083481739602,
+ "loss": 3.9027457237243652,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.009498288854956627,
+ "learning_rate": 0.0010510419539845777,
+ "loss": 3.917316436767578,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.00849096104502678,
+ "learning_rate": 0.001049775933597813,
+ "loss": 3.9229235649108887,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.008343606255948544,
+ "learning_rate": 0.0010485102880336747,
+ "loss": 3.9138073921203613,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.010342134162783623,
+ "learning_rate": 0.0010472450183118727,
+ "loss": 3.8932929039001465,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.010298306122422218,
+ "learning_rate": 0.00104598012545181,
+ "loss": 3.9112677574157715,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.00827560480684042,
+ "learning_rate": 0.001044715610472589,
+ "loss": 3.894556760787964,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.008993102237582207,
+ "learning_rate": 0.0010434514743930064,
+ "loss": 3.905391216278076,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.008632538840174675,
+ "learning_rate": 0.0010421877182315536,
+ "loss": 3.896714210510254,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.008542424067854881,
+ "learning_rate": 0.0010409243430064162,
+ "loss": 3.9145569801330566,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.008970450609922409,
+ "learning_rate": 0.0010396613497354725,
+ "loss": 3.9170360565185547,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.008744142949581146,
+ "learning_rate": 0.0010383987394362942,
+ "loss": 3.927412509918213,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.008229847066104412,
+ "learning_rate": 0.0010371365131261424,
+ "loss": 3.928352117538452,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.00818693358451128,
+ "learning_rate": 0.0010358746718219717,
+ "loss": 3.9219343662261963,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.008848588913679123,
+ "learning_rate": 0.001034613216540423,
+ "loss": 3.9346699714660645,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.00834780652076006,
+ "learning_rate": 0.0010333521482978295,
+ "loss": 3.9144492149353027,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.00940527394413948,
+ "learning_rate": 0.0010320914681102098,
+ "loss": 3.951573371887207,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.008884521201252937,
+ "learning_rate": 0.0010308311769932725,
+ "loss": 3.899625778198242,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.007863261736929417,
+ "learning_rate": 0.0010295712759624107,
+ "loss": 3.9321727752685547,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.008651153184473515,
+ "learning_rate": 0.0010283117660327035,
+ "loss": 3.9030094146728516,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.00822074618190527,
+ "learning_rate": 0.0010270526482189163,
+ "loss": 3.939466953277588,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.008701153099536896,
+ "learning_rate": 0.0010257939235354964,
+ "loss": 3.9099626541137695,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.008418620564043522,
+ "learning_rate": 0.0010245355929965774,
+ "loss": 3.914116382598877,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.008339700289070606,
+ "learning_rate": 0.001023277657615971,
+ "loss": 3.9111878871917725,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.008304647170007229,
+ "learning_rate": 0.0010220201184071747,
+ "loss": 3.950974464416504,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.008436229079961777,
+ "learning_rate": 0.001020762976383365,
+ "loss": 3.9631457328796387,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.0074487063102424145,
+ "learning_rate": 0.001019506232557398,
+ "loss": 3.902984619140625,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.008736515417695045,
+ "learning_rate": 0.0010182498879418096,
+ "loss": 3.9268949031829834,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.009666468016803265,
+ "learning_rate": 0.0010169939435488145,
+ "loss": 3.9274301528930664,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.395970821380615,
+ "eval_runtime": 53.7848,
+ "eval_samples_per_second": 45.403,
+ "eval_steps_per_second": 1.432,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.009094545617699623,
+ "learning_rate": 0.001015738400390304,
+ "loss": 3.866425037384033,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.009273733012378216,
+ "learning_rate": 0.0010144832594778462,
+ "loss": 3.8569841384887695,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.009205271489918232,
+ "learning_rate": 0.0010132285218226855,
+ "loss": 3.8981378078460693,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.009963045828044415,
+ "learning_rate": 0.0010119741884357424,
+ "loss": 3.881542444229126,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.008536554872989655,
+ "learning_rate": 0.0010107202603276096,
+ "loss": 3.851253032684326,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.01074561569839716,
+ "learning_rate": 0.001009466738508555,
+ "loss": 3.8731184005737305,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.010809902101755142,
+ "learning_rate": 0.0010082136239885177,
+ "loss": 3.863889217376709,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.008289352990686893,
+ "learning_rate": 0.0010069609177771103,
+ "loss": 3.839066743850708,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.00792208593338728,
+ "learning_rate": 0.001005708620883615,
+ "loss": 3.8634843826293945,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.008381598629057407,
+ "learning_rate": 0.0010044567343169845,
+ "loss": 3.868283987045288,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.009784617461264133,
+ "learning_rate": 0.0010032052590858425,
+ "loss": 3.853074550628662,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.009462198242545128,
+ "learning_rate": 0.0010019541961984782,
+ "loss": 3.889979600906372,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.010207819752395153,
+ "learning_rate": 0.001000703546662852,
+ "loss": 3.8636741638183594,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.009123551659286022,
+ "learning_rate": 0.0009994533114865887,
+ "loss": 3.8578555583953857,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.009006140753626823,
+ "learning_rate": 0.0009982034916769806,
+ "loss": 3.8828015327453613,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.0091114342212677,
+ "learning_rate": 0.0009969540882409846,
+ "loss": 3.880486011505127,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.008851367980241776,
+ "learning_rate": 0.0009957051021852227,
+ "loss": 3.8618125915527344,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.008196741342544556,
+ "learning_rate": 0.0009944565345159806,
+ "loss": 3.8314175605773926,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.00863513071089983,
+ "learning_rate": 0.000993208386239206,
+ "loss": 3.856212615966797,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.008624681271612644,
+ "learning_rate": 0.0009919606583605107,
+ "loss": 3.868927001953125,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.009208842180669308,
+ "learning_rate": 0.0009907133518851647,
+ "loss": 3.9072189331054688,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.010200913064181805,
+ "learning_rate": 0.0009894664678181027,
+ "loss": 3.875272750854492,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.009892710484564304,
+ "learning_rate": 0.000988220007163914,
+ "loss": 3.888686418533325,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.009873585775494576,
+ "learning_rate": 0.000986973970926851,
+ "loss": 3.8607118129730225,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.009322221390902996,
+ "learning_rate": 0.0009857283601108228,
+ "loss": 3.8852052688598633,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.008141528815031052,
+ "learning_rate": 0.0009844831757193947,
+ "loss": 3.888213872909546,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.008191613480448723,
+ "learning_rate": 0.00098323841875579,
+ "loss": 3.9073967933654785,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.008419081568717957,
+ "learning_rate": 0.0009819940902228864,
+ "loss": 3.879223346710205,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.009037869982421398,
+ "learning_rate": 0.0009807501911232171,
+ "loss": 3.883303165435791,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.00854078121483326,
+ "learning_rate": 0.0009795067224589687,
+ "loss": 3.926767587661743,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.008590771816670895,
+ "learning_rate": 0.0009782636852319825,
+ "loss": 3.913999557495117,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.00859815627336502,
+ "learning_rate": 0.0009770210804437498,
+ "loss": 3.8828063011169434,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.008970260620117188,
+ "learning_rate": 0.0009757789090954157,
+ "loss": 3.9174866676330566,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.00877175759524107,
+ "learning_rate": 0.0009745371721877754,
+ "loss": 3.8879401683807373,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.008541764691472054,
+ "learning_rate": 0.0009732958707212731,
+ "loss": 3.8717427253723145,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.008466241881251335,
+ "learning_rate": 0.0009720550056960039,
+ "loss": 3.8944485187530518,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.00808491837233305,
+ "learning_rate": 0.0009708145781117096,
+ "loss": 3.8617026805877686,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.00842207670211792,
+ "learning_rate": 0.0009695745889677805,
+ "loss": 3.887824535369873,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.00830599945038557,
+ "learning_rate": 0.0009683350392632533,
+ "loss": 3.8612070083618164,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.008417914621531963,
+ "learning_rate": 0.0009670959299968111,
+ "loss": 3.880140781402588,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.008542059920728207,
+ "learning_rate": 0.0009658572621667822,
+ "loss": 3.8951029777526855,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.008822236210107803,
+ "learning_rate": 0.000964619036771138,
+ "loss": 3.8962583541870117,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.009151184931397438,
+ "learning_rate": 0.0009633812548074952,
+ "loss": 3.882500648498535,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.007803036365658045,
+ "learning_rate": 0.0009621439172731115,
+ "loss": 3.8740782737731934,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.009350850246846676,
+ "learning_rate": 0.0009609070251648884,
+ "loss": 3.8627419471740723,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.008319144137203693,
+ "learning_rate": 0.0009596705794793667,
+ "loss": 3.9084486961364746,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.009198528714478016,
+ "learning_rate": 0.0009584345812127282,
+ "loss": 3.92141056060791,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.010480938479304314,
+ "learning_rate": 0.0009571990313607958,
+ "loss": 3.8916103839874268,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.01045288611203432,
+ "learning_rate": 0.0009559639309190278,
+ "loss": 3.8863210678100586,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.009121391922235489,
+ "learning_rate": 0.0009547292808825239,
+ "loss": 3.9167044162750244,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.01093223039060831,
+ "learning_rate": 0.0009534950822460184,
+ "loss": 3.9046711921691895,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.01214868575334549,
+ "learning_rate": 0.0009522613360038835,
+ "loss": 3.9126768112182617,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.01055280864238739,
+ "learning_rate": 0.0009510280431501255,
+ "loss": 3.9214353561401367,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.009021313861012459,
+ "learning_rate": 0.0009497952046783867,
+ "loss": 3.8795580863952637,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.00863752793520689,
+ "learning_rate": 0.0009485628215819428,
+ "loss": 3.893232822418213,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.009327626787126064,
+ "learning_rate": 0.0009473308948537021,
+ "loss": 3.917377471923828,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.01003898587077856,
+ "learning_rate": 0.0009460994254862066,
+ "loss": 3.8818931579589844,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.00892532430589199,
+ "learning_rate": 0.0009448684144716278,
+ "loss": 3.8761978149414062,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.008396463468670845,
+ "learning_rate": 0.0009436378628017702,
+ "loss": 3.8665928840637207,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.00784370768815279,
+ "learning_rate": 0.0009424077714680662,
+ "loss": 3.8850138187408447,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.00809321366250515,
+ "learning_rate": 0.0009411781414615784,
+ "loss": 3.917078733444214,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.008685288950800896,
+ "learning_rate": 0.000939948973772998,
+ "loss": 3.8865013122558594,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.007291820365935564,
+ "learning_rate": 0.0009387202693926425,
+ "loss": 3.8967666625976562,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.008459319360554218,
+ "learning_rate": 0.0009374920293104573,
+ "loss": 3.9043235778808594,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.007924137637019157,
+ "learning_rate": 0.0009362642545160123,
+ "loss": 3.9216856956481934,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.007874754257500172,
+ "learning_rate": 0.0009350369459985052,
+ "loss": 3.879725933074951,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.008643723092973232,
+ "learning_rate": 0.0009338101047467544,
+ "loss": 3.9303886890411377,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.008451704867184162,
+ "learning_rate": 0.000932583731749205,
+ "loss": 3.9352827072143555,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.008137754164636135,
+ "learning_rate": 0.0009313578279939236,
+ "loss": 3.930800437927246,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.009277567267417908,
+ "learning_rate": 0.0009301323944685979,
+ "loss": 3.9194300174713135,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.009428970515727997,
+ "learning_rate": 0.0009289074321605383,
+ "loss": 3.874298572540283,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.0089727733284235,
+ "learning_rate": 0.0009276829420566743,
+ "loss": 3.893770694732666,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.388528823852539,
+ "eval_runtime": 51.5859,
+ "eval_samples_per_second": 47.338,
+ "eval_steps_per_second": 1.493,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.009470812045037746,
+ "learning_rate": 0.000926458925143556,
+ "loss": 3.8514318466186523,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.011485518887639046,
+ "learning_rate": 0.0009252353824073511,
+ "loss": 3.857734203338623,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.010095939971506596,
+ "learning_rate": 0.0009240123148338457,
+ "loss": 3.819117546081543,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.010103728622198105,
+ "learning_rate": 0.0009227897234084444,
+ "loss": 3.842660427093506,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.01148147787898779,
+ "learning_rate": 0.0009215676091161661,
+ "loss": 3.821089029312134,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.012331553734838963,
+ "learning_rate": 0.0009203459729416467,
+ "loss": 3.8194339275360107,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.011274794116616249,
+ "learning_rate": 0.0009191248158691358,
+ "loss": 3.8839240074157715,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.009603365324437618,
+ "learning_rate": 0.0009179041388824986,
+ "loss": 3.806617259979248,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.010824795812368393,
+ "learning_rate": 0.0009166839429652112,
+ "loss": 3.823228597640991,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.01014280878007412,
+ "learning_rate": 0.0009154642291003637,
+ "loss": 3.860605239868164,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.009052124805748463,
+ "learning_rate": 0.0009142449982706591,
+ "loss": 3.84226131439209,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.009465130977332592,
+ "learning_rate": 0.0009130262514584074,
+ "loss": 3.8674912452697754,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.009991783648729324,
+ "learning_rate": 0.0009118079896455326,
+ "loss": 3.8775739669799805,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.009288253262639046,
+ "learning_rate": 0.0009105902138135652,
+ "loss": 3.7978098392486572,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.009964323602616787,
+ "learning_rate": 0.0009093729249436459,
+ "loss": 3.8362321853637695,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.010977157391607761,
+ "learning_rate": 0.0009081561240165218,
+ "loss": 3.857304096221924,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.009241319261491299,
+ "learning_rate": 0.0009069398120125479,
+ "loss": 3.8659310340881348,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.009211038239300251,
+ "learning_rate": 0.0009057239899116844,
+ "loss": 3.8764538764953613,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.009649050422012806,
+ "learning_rate": 0.000904508658693497,
+ "loss": 3.8702778816223145,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.00954380165785551,
+ "learning_rate": 0.0009032938193371573,
+ "loss": 3.8273327350616455,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.008831270970404148,
+ "learning_rate": 0.0009020794728214378,
+ "loss": 3.821664810180664,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.008367286063730717,
+ "learning_rate": 0.000900865620124717,
+ "loss": 3.8586034774780273,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.008805174380540848,
+ "learning_rate": 0.0008996522622249732,
+ "loss": 3.811522960662842,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.00873226672410965,
+ "learning_rate": 0.0008984394000997873,
+ "loss": 3.8538150787353516,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.008930203504860401,
+ "learning_rate": 0.0008972270347263402,
+ "loss": 3.8448486328125,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.009300434030592442,
+ "learning_rate": 0.0008960151670814125,
+ "loss": 3.842113494873047,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.009598663076758385,
+ "learning_rate": 0.0008948037981413855,
+ "loss": 3.886852741241455,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.008400370366871357,
+ "learning_rate": 0.0008935929288822357,
+ "loss": 3.852076292037964,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.00891344528645277,
+ "learning_rate": 0.0008923825602795391,
+ "loss": 3.8480606079101562,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.00869396049529314,
+ "learning_rate": 0.0008911726933084678,
+ "loss": 3.8164706230163574,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.007914266549050808,
+ "learning_rate": 0.0008899633289437904,
+ "loss": 3.8768272399902344,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.008398533798754215,
+ "learning_rate": 0.0008887544681598687,
+ "loss": 3.8625502586364746,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.00887146033346653,
+ "learning_rate": 0.0008875461119306601,
+ "loss": 3.8934593200683594,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.008642833679914474,
+ "learning_rate": 0.0008863382612297172,
+ "loss": 3.8391976356506348,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.008099788799881935,
+ "learning_rate": 0.0008851309170301816,
+ "loss": 3.827965497970581,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.008166342042386532,
+ "learning_rate": 0.0008839240803047894,
+ "loss": 3.8999152183532715,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.009095479734241962,
+ "learning_rate": 0.0008827177520258671,
+ "loss": 3.876927614212036,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.009246554225683212,
+ "learning_rate": 0.0008815119331653321,
+ "loss": 3.8692450523376465,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.007059223018586636,
+ "learning_rate": 0.0008803066246946898,
+ "loss": 3.869400978088379,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.008499744348227978,
+ "learning_rate": 0.0008791018275850363,
+ "loss": 3.8303189277648926,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.007667839527130127,
+ "learning_rate": 0.0008778975428070543,
+ "loss": 3.827998638153076,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.00765089550986886,
+ "learning_rate": 0.0008766937713310145,
+ "loss": 3.8554482460021973,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.00862938817590475,
+ "learning_rate": 0.0008754905141267749,
+ "loss": 3.905458927154541,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.008733712136745453,
+ "learning_rate": 0.0008742877721637758,
+ "loss": 3.8564133644104004,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.0071738120168447495,
+ "learning_rate": 0.0008730855464110471,
+ "loss": 3.878807544708252,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.008418241515755653,
+ "learning_rate": 0.0008718838378371989,
+ "loss": 3.8399720191955566,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.008526502177119255,
+ "learning_rate": 0.0008706826474104263,
+ "loss": 3.847656488418579,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.00732462527230382,
+ "learning_rate": 0.0008694819760985069,
+ "loss": 3.8839492797851562,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.008424196392297745,
+ "learning_rate": 0.0008682818248688003,
+ "loss": 3.861239433288574,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.008984955959022045,
+ "learning_rate": 0.0008670821946882465,
+ "loss": 3.856529712677002,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.009004034101963043,
+ "learning_rate": 0.000865883086523365,
+ "loss": 3.8514528274536133,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.009097260423004627,
+ "learning_rate": 0.0008646845013402563,
+ "loss": 3.8890581130981445,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.009244748391211033,
+ "learning_rate": 0.0008634864401045988,
+ "loss": 3.872745990753174,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.00788815412670374,
+ "learning_rate": 0.0008622889037816488,
+ "loss": 3.8777575492858887,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.008943610824644566,
+ "learning_rate": 0.0008610918933362396,
+ "loss": 3.8825736045837402,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.008154108189046383,
+ "learning_rate": 0.0008598954097327809,
+ "loss": 3.877514362335205,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.008689315989613533,
+ "learning_rate": 0.0008586994539352587,
+ "loss": 3.854720115661621,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.00904505793005228,
+ "learning_rate": 0.0008575040269072315,
+ "loss": 3.911463499069214,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.007773210760205984,
+ "learning_rate": 0.0008563091296118342,
+ "loss": 3.875370740890503,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.008437697775661945,
+ "learning_rate": 0.0008551147630117737,
+ "loss": 3.8457469940185547,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.009780957363545895,
+ "learning_rate": 0.0008539209280693296,
+ "loss": 3.890479564666748,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.007929213345050812,
+ "learning_rate": 0.0008527276257463532,
+ "loss": 3.876030445098877,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.009121386334300041,
+ "learning_rate": 0.0008515348570042669,
+ "loss": 3.9252734184265137,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.00981531199067831,
+ "learning_rate": 0.000850342622804063,
+ "loss": 3.8997983932495117,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.009150844067335129,
+ "learning_rate": 0.0008491509241063025,
+ "loss": 3.85115385055542,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.00992978923022747,
+ "learning_rate": 0.0008479597618711154,
+ "loss": 3.9138102531433105,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.010906689800322056,
+ "learning_rate": 0.0008467691370582001,
+ "loss": 3.892329692840576,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.008644519373774529,
+ "learning_rate": 0.0008455790506268213,
+ "loss": 3.9071311950683594,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.009030954912304878,
+ "learning_rate": 0.0008443895035358111,
+ "loss": 3.867384910583496,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.010806960985064507,
+ "learning_rate": 0.0008432004967435638,
+ "loss": 3.877983331680298,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.011148087680339813,
+ "learning_rate": 0.0008420120312080436,
+ "loss": 3.8795924186706543,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01046178862452507,
+ "learning_rate": 0.0008408241078867731,
+ "loss": 3.887709617614746,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.383835792541504,
+ "eval_runtime": 53.0778,
+ "eval_samples_per_second": 46.008,
+ "eval_steps_per_second": 1.451,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.009721867740154266,
+ "learning_rate": 0.0008396367277368421,
+ "loss": 3.8261327743530273,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.01091420091688633,
+ "learning_rate": 0.000838449891714901,
+ "loss": 3.7837276458740234,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.010423852130770683,
+ "learning_rate": 0.0008372636007771623,
+ "loss": 3.796450138092041,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.011928882449865341,
+ "learning_rate": 0.0008360778558793996,
+ "loss": 3.794717311859131,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.010779444128274918,
+ "learning_rate": 0.000834892657976945,
+ "loss": 3.8477983474731445,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.009554282762110233,
+ "learning_rate": 0.0008337080080246915,
+ "loss": 3.793790817260742,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.009768152609467506,
+ "learning_rate": 0.0008325239069770906,
+ "loss": 3.860384225845337,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.010801832191646099,
+ "learning_rate": 0.0008313403557881513,
+ "loss": 3.816796064376831,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.01046004518866539,
+ "learning_rate": 0.0008301573554114378,
+ "loss": 3.8064301013946533,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.009319447912275791,
+ "learning_rate": 0.0008289749068000744,
+ "loss": 3.832408905029297,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01011407095938921,
+ "learning_rate": 0.000827793010906737,
+ "loss": 3.8276009559631348,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.010393989272415638,
+ "learning_rate": 0.0008266116686836582,
+ "loss": 3.83754825592041,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.00883004255592823,
+ "learning_rate": 0.0008254308810826244,
+ "loss": 3.8315513134002686,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.008549787104129791,
+ "learning_rate": 0.0008242506490549745,
+ "loss": 3.792919158935547,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.009668390266597271,
+ "learning_rate": 0.0008230709735516006,
+ "loss": 3.8110954761505127,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.009008699096739292,
+ "learning_rate": 0.0008218918555229445,
+ "loss": 3.8598508834838867,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.008660725317895412,
+ "learning_rate": 0.0008207132959190026,
+ "loss": 3.839632511138916,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.00827972125262022,
+ "learning_rate": 0.0008195352956893172,
+ "loss": 3.8428683280944824,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.008117233403027058,
+ "learning_rate": 0.0008183578557829825,
+ "loss": 3.8434956073760986,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.008314455859363079,
+ "learning_rate": 0.0008171809771486405,
+ "loss": 3.8079686164855957,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.00865850131958723,
+ "learning_rate": 0.0008160046607344811,
+ "loss": 3.846999406814575,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.0096864253282547,
+ "learning_rate": 0.0008148289074882416,
+ "loss": 3.8227243423461914,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.00840236246585846,
+ "learning_rate": 0.0008136537183572045,
+ "loss": 3.8400344848632812,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.007451791316270828,
+ "learning_rate": 0.0008124790942881984,
+ "loss": 3.8318095207214355,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.00858070608228445,
+ "learning_rate": 0.000811305036227597,
+ "loss": 3.8232922554016113,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.00765061192214489,
+ "learning_rate": 0.0008101315451213178,
+ "loss": 3.81624174118042,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.009304090403020382,
+ "learning_rate": 0.0008089586219148213,
+ "loss": 3.8715004920959473,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.009136030450463295,
+ "learning_rate": 0.0008077862675531103,
+ "loss": 3.825037956237793,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.007460939697921276,
+ "learning_rate": 0.0008066144829807303,
+ "loss": 3.811749219894409,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.007510480936616659,
+ "learning_rate": 0.0008054432691417659,
+ "loss": 3.8313679695129395,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.008825462311506271,
+ "learning_rate": 0.0008042726269798433,
+ "loss": 3.83119535446167,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.009002318605780602,
+ "learning_rate": 0.0008031025574381281,
+ "loss": 3.8542468547821045,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.009266122244298458,
+ "learning_rate": 0.0008019330614593237,
+ "loss": 3.8328065872192383,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.008935115300118923,
+ "learning_rate": 0.0008007641399856727,
+ "loss": 3.8198719024658203,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.008021850138902664,
+ "learning_rate": 0.0007995957939589526,
+ "loss": 3.828382968902588,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.008035171777009964,
+ "learning_rate": 0.0007984280243204805,
+ "loss": 3.879868507385254,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.008553579449653625,
+ "learning_rate": 0.0007972608320111059,
+ "loss": 3.842162609100342,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.008016746491193771,
+ "learning_rate": 0.000796094217971215,
+ "loss": 3.85674786567688,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.007781418971717358,
+ "learning_rate": 0.0007949281831407276,
+ "loss": 3.7968859672546387,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.008552255108952522,
+ "learning_rate": 0.0007937627284590971,
+ "loss": 3.8305506706237793,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.007390565704554319,
+ "learning_rate": 0.0007925978548653096,
+ "loss": 3.8228843212127686,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.007611850742250681,
+ "learning_rate": 0.0007914335632978812,
+ "loss": 3.868957996368408,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.008188549429178238,
+ "learning_rate": 0.0007902698546948628,
+ "loss": 3.836078405380249,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.008198797702789307,
+ "learning_rate": 0.0007891067299938318,
+ "loss": 3.8637044429779053,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.008353965356945992,
+ "learning_rate": 0.0007879441901318977,
+ "loss": 3.833843946456909,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.0086967209354043,
+ "learning_rate": 0.0007867822360456963,
+ "loss": 3.850536823272705,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.009151408448815346,
+ "learning_rate": 0.0007856208686713949,
+ "loss": 3.8453407287597656,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.008010942488908768,
+ "learning_rate": 0.0007844600889446858,
+ "loss": 3.8679513931274414,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.007813206873834133,
+ "learning_rate": 0.0007832998978007873,
+ "loss": 3.860609531402588,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.007759213913232088,
+ "learning_rate": 0.0007821402961744448,
+ "loss": 3.803863048553467,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.008045864291489124,
+ "learning_rate": 0.0007809812849999293,
+ "loss": 3.858161449432373,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.008212368935346603,
+ "learning_rate": 0.0007798228652110347,
+ "loss": 3.8311336040496826,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.007982932962477207,
+ "learning_rate": 0.0007786650377410774,
+ "loss": 3.870339870452881,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.0071402136236429214,
+ "learning_rate": 0.0007775078035229002,
+ "loss": 3.7922263145446777,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.008564750663936138,
+ "learning_rate": 0.0007763511634888656,
+ "loss": 3.83981990814209,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.008078585378825665,
+ "learning_rate": 0.0007751951185708562,
+ "loss": 3.852938652038574,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.00801754929125309,
+ "learning_rate": 0.0007740396697002774,
+ "loss": 3.8434009552001953,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.008973612450063229,
+ "learning_rate": 0.0007728848178080535,
+ "loss": 3.8238563537597656,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.010199865326285362,
+ "learning_rate": 0.0007717305638246281,
+ "loss": 3.8891735076904297,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.009620471857488155,
+ "learning_rate": 0.0007705769086799617,
+ "loss": 3.8611769676208496,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.007981563918292522,
+ "learning_rate": 0.0007694238533035337,
+ "loss": 3.8496928215026855,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.008856567554175854,
+ "learning_rate": 0.0007682713986243415,
+ "loss": 3.8366098403930664,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.008323253132402897,
+ "learning_rate": 0.0007671195455708952,
+ "loss": 3.8224244117736816,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.00814893003553152,
+ "learning_rate": 0.0007659682950712228,
+ "loss": 3.8611154556274414,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.009146672673523426,
+ "learning_rate": 0.0007648176480528659,
+ "loss": 3.8258843421936035,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.007573423441499472,
+ "learning_rate": 0.0007636676054428805,
+ "loss": 3.888007640838623,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.00813984964042902,
+ "learning_rate": 0.0007625181681678343,
+ "loss": 3.860306739807129,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.007936778478324413,
+ "learning_rate": 0.0007613693371538078,
+ "loss": 3.8808555603027344,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.008434669114649296,
+ "learning_rate": 0.0007602211133263956,
+ "loss": 3.854154348373413,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.007857564836740494,
+ "learning_rate": 0.0007590734976106985,
+ "loss": 3.8240737915039062,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.007989327423274517,
+ "learning_rate": 0.0007579264909313315,
+ "loss": 3.850971221923828,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.008888393640518188,
+ "learning_rate": 0.0007567800942124152,
+ "loss": 3.862267255783081,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.38022518157959,
+ "eval_runtime": 51.4686,
+ "eval_samples_per_second": 47.446,
+ "eval_steps_per_second": 1.496,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.00952722318470478,
+ "learning_rate": 0.0007556343083775832,
+ "loss": 3.7705495357513428,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01099323108792305,
+ "learning_rate": 0.0007544891343499723,
+ "loss": 3.80812406539917,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.010298571549355984,
+ "learning_rate": 0.0007533445730522295,
+ "loss": 3.782794237136841,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.008139589801430702,
+ "learning_rate": 0.000752200625406507,
+ "loss": 3.7877392768859863,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.009031252935528755,
+ "learning_rate": 0.0007510572923344628,
+ "loss": 3.8181564807891846,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.00942459236830473,
+ "learning_rate": 0.0007499145747572603,
+ "loss": 3.807425022125244,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.00922947097569704,
+ "learning_rate": 0.0007487724735955648,
+ "loss": 3.7923264503479004,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.009551790542900562,
+ "learning_rate": 0.0007476309897695485,
+ "loss": 3.790083885192871,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.008996761403977871,
+ "learning_rate": 0.0007464901241988831,
+ "loss": 3.836491107940674,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.009766182862222195,
+ "learning_rate": 0.0007453498778027433,
+ "loss": 3.7829647064208984,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.007708575110882521,
+ "learning_rate": 0.0007442102514998058,
+ "loss": 3.803400993347168,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.008606902323663235,
+ "learning_rate": 0.0007430712462082464,
+ "loss": 3.810825824737549,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.009462927468121052,
+ "learning_rate": 0.0007419328628457416,
+ "loss": 3.784748077392578,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.010283366777002811,
+ "learning_rate": 0.0007407951023294655,
+ "loss": 3.744312286376953,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.008952794596552849,
+ "learning_rate": 0.0007396579655760917,
+ "loss": 3.8098015785217285,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.008446207270026207,
+ "learning_rate": 0.0007385214535017907,
+ "loss": 3.8306760787963867,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.007988285273313522,
+ "learning_rate": 0.0007373855670222302,
+ "loss": 3.7976696491241455,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.007586437277495861,
+ "learning_rate": 0.0007362503070525718,
+ "loss": 3.785891056060791,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.00775605533272028,
+ "learning_rate": 0.0007351156745074757,
+ "loss": 3.8176379203796387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.007727691903710365,
+ "learning_rate": 0.0007339816703010948,
+ "loss": 3.791743278503418,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.008311517536640167,
+ "learning_rate": 0.0007328482953470747,
+ "loss": 3.818769693374634,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.007610874716192484,
+ "learning_rate": 0.0007317155505585556,
+ "loss": 3.76772141456604,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.0078107016161084175,
+ "learning_rate": 0.0007305834368481699,
+ "loss": 3.822758197784424,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.006906864698976278,
+ "learning_rate": 0.0007294519551280412,
+ "loss": 3.832834005355835,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.008230660110712051,
+ "learning_rate": 0.0007283211063097829,
+ "loss": 3.85087513923645,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.008201118558645248,
+ "learning_rate": 0.0007271908913044997,
+ "loss": 3.7777247428894043,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.0070146857760846615,
+ "learning_rate": 0.000726061311022787,
+ "loss": 3.837517738342285,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.0076036700047552586,
+ "learning_rate": 0.0007249323663747254,
+ "loss": 3.8358592987060547,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.007238952908664942,
+ "learning_rate": 0.0007238040582698858,
+ "loss": 3.828089952468872,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.00825269054621458,
+ "learning_rate": 0.0007226763876173259,
+ "loss": 3.806016445159912,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.008368568494915962,
+ "learning_rate": 0.0007215493553255898,
+ "loss": 3.8163394927978516,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.00728471577167511,
+ "learning_rate": 0.0007204229623027063,
+ "loss": 3.835597038269043,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.00784830842167139,
+ "learning_rate": 0.0007192972094561897,
+ "loss": 3.8050971031188965,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.007714141625910997,
+ "learning_rate": 0.0007181720976930404,
+ "loss": 3.8031606674194336,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.009128103032708168,
+ "learning_rate": 0.000717047627919739,
+ "loss": 3.77107572555542,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.008033272810280323,
+ "learning_rate": 0.000715923801042251,
+ "loss": 3.8615059852600098,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.007732565049082041,
+ "learning_rate": 0.0007148006179660235,
+ "loss": 3.8277132511138916,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.007382966112345457,
+ "learning_rate": 0.0007136780795959853,
+ "loss": 3.803431987762451,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.007517020218074322,
+ "learning_rate": 0.0007125561868365442,
+ "loss": 3.799734354019165,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.008041434921324253,
+ "learning_rate": 0.0007114349405915895,
+ "loss": 3.7848258018493652,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.007875649258494377,
+ "learning_rate": 0.0007103143417644891,
+ "loss": 3.8136563301086426,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.008213640190660954,
+ "learning_rate": 0.000709194391258089,
+ "loss": 3.8518495559692383,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.00801054760813713,
+ "learning_rate": 0.0007080750899747139,
+ "loss": 3.7956271171569824,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.008064007386565208,
+ "learning_rate": 0.0007069564388161631,
+ "loss": 3.854595184326172,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.008300521411001682,
+ "learning_rate": 0.0007058384386837154,
+ "loss": 3.8203625679016113,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.008640496991574764,
+ "learning_rate": 0.000704721090478122,
+ "loss": 3.793574810028076,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.007684200536459684,
+ "learning_rate": 0.0007036043950996105,
+ "loss": 3.8564438819885254,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.008431905880570412,
+ "learning_rate": 0.0007024883534478823,
+ "loss": 3.8271119594573975,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.008137186989188194,
+ "learning_rate": 0.0007013729664221122,
+ "loss": 3.8219337463378906,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.0076744635589420795,
+ "learning_rate": 0.0007002582349209474,
+ "loss": 3.8353049755096436,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.008388740941882133,
+ "learning_rate": 0.0006991441598425065,
+ "loss": 3.795104503631592,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.007856911979615688,
+ "learning_rate": 0.0006980307420843798,
+ "loss": 3.818411111831665,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.008048887364566326,
+ "learning_rate": 0.000696917982543628,
+ "loss": 3.86599063873291,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.007949979975819588,
+ "learning_rate": 0.0006958058821167812,
+ "loss": 3.8026742935180664,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.008162821643054485,
+ "learning_rate": 0.0006946944416998389,
+ "loss": 3.833751678466797,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.008190765976905823,
+ "learning_rate": 0.0006935836621882682,
+ "loss": 3.8164236545562744,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.009714310988783836,
+ "learning_rate": 0.000692473544477005,
+ "loss": 3.8367366790771484,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.008075987920165062,
+ "learning_rate": 0.0006913640894604498,
+ "loss": 3.8108773231506348,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.009075167588889599,
+ "learning_rate": 0.0006902552980324713,
+ "loss": 3.8444485664367676,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.008678440935909748,
+ "learning_rate": 0.0006891471710864022,
+ "loss": 3.8384461402893066,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.007705411408096552,
+ "learning_rate": 0.0006880397095150414,
+ "loss": 3.825443744659424,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.00782099086791277,
+ "learning_rate": 0.0006869329142106488,
+ "loss": 3.8331687450408936,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.007542501203715801,
+ "learning_rate": 0.0006858267860649509,
+ "loss": 3.7905991077423096,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.008315361104905605,
+ "learning_rate": 0.0006847213259691356,
+ "loss": 3.861527442932129,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.008529976941645145,
+ "learning_rate": 0.0006836165348138508,
+ "loss": 3.8332972526550293,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.008117211051285267,
+ "learning_rate": 0.0006825124134892072,
+ "loss": 3.7961089611053467,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.008409497328102589,
+ "learning_rate": 0.0006814089628847755,
+ "loss": 3.812218189239502,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.007592010777443647,
+ "learning_rate": 0.0006803061838895864,
+ "loss": 3.801304340362549,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.007820053026080132,
+ "learning_rate": 0.0006792040773921283,
+ "loss": 3.8517117500305176,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.0073044621385633945,
+ "learning_rate": 0.0006781026442803479,
+ "loss": 3.8253560066223145,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.007546546868979931,
+ "learning_rate": 0.0006770018854416522,
+ "loss": 3.800802230834961,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.008878474123775959,
+ "learning_rate": 0.0006759018017629007,
+ "loss": 3.8133859634399414,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.375829219818115,
+ "eval_runtime": 51.3756,
+ "eval_samples_per_second": 47.532,
+ "eval_steps_per_second": 1.499,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.00814585480839014,
+ "learning_rate": 0.0006748023941304115,
+ "loss": 3.7970919609069824,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.00824262760579586,
+ "learning_rate": 0.0006737036634299577,
+ "loss": 3.750131607055664,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.008675593882799149,
+ "learning_rate": 0.000672605610546767,
+ "loss": 3.791456699371338,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.009052547626197338,
+ "learning_rate": 0.00067150823636552,
+ "loss": 3.7762060165405273,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.00818016566336155,
+ "learning_rate": 0.0006704115417703514,
+ "loss": 3.750631332397461,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.00800115056335926,
+ "learning_rate": 0.0006693155276448489,
+ "loss": 3.811519145965576,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.009032433852553368,
+ "learning_rate": 0.0006682201948720509,
+ "loss": 3.7805609703063965,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.0077886078506708145,
+ "learning_rate": 0.0006671255443344474,
+ "loss": 3.749366521835327,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.008603204973042011,
+ "learning_rate": 0.0006660315769139774,
+ "loss": 3.7686312198638916,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.008361323736608028,
+ "learning_rate": 0.0006649382934920329,
+ "loss": 3.740269660949707,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.00948982872068882,
+ "learning_rate": 0.0006638456949494507,
+ "loss": 3.7874045372009277,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.008958354592323303,
+ "learning_rate": 0.0006627537821665185,
+ "loss": 3.78987455368042,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.008702470920979977,
+ "learning_rate": 0.0006616625560229704,
+ "loss": 3.751061201095581,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.008424391970038414,
+ "learning_rate": 0.000660572017397988,
+ "loss": 3.80816650390625,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.009022071026265621,
+ "learning_rate": 0.0006594821671701989,
+ "loss": 3.726067304611206,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.007974586449563503,
+ "learning_rate": 0.0006583930062176739,
+ "loss": 3.78564453125,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.007887622341513634,
+ "learning_rate": 0.0006573045354179329,
+ "loss": 3.7993059158325195,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.007782666943967342,
+ "learning_rate": 0.0006562167556479352,
+ "loss": 3.7440507411956787,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.007923880591988564,
+ "learning_rate": 0.000655129667784086,
+ "loss": 3.7920784950256348,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.007850701920688152,
+ "learning_rate": 0.0006540432727022325,
+ "loss": 3.7833633422851562,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.007855751551687717,
+ "learning_rate": 0.0006529575712776633,
+ "loss": 3.812349319458008,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.007832377217710018,
+ "learning_rate": 0.0006518725643851091,
+ "loss": 3.7815146446228027,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.006997724529355764,
+ "learning_rate": 0.0006507882528987395,
+ "loss": 3.795346975326538,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.007433670572936535,
+ "learning_rate": 0.0006497046376921648,
+ "loss": 3.841017246246338,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.0075767990201711655,
+ "learning_rate": 0.0006486217196384344,
+ "loss": 3.7913427352905273,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.007551983464509249,
+ "learning_rate": 0.0006475394996100359,
+ "loss": 3.765105962753296,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.007235546596348286,
+ "learning_rate": 0.0006464579784788943,
+ "loss": 3.7668676376342773,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.007872921414673328,
+ "learning_rate": 0.0006453771571163716,
+ "loss": 3.770536422729492,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.007598782889544964,
+ "learning_rate": 0.0006442970363932667,
+ "loss": 3.796818733215332,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.007902889512479305,
+ "learning_rate": 0.0006432176171798124,
+ "loss": 3.824531078338623,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.008964902721345425,
+ "learning_rate": 0.0006421389003456776,
+ "loss": 3.781778335571289,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.009936625137925148,
+ "learning_rate": 0.000641060886759965,
+ "loss": 3.750248670578003,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.008381214924156666,
+ "learning_rate": 0.0006399835772912106,
+ "loss": 3.797954797744751,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.007752922363579273,
+ "learning_rate": 0.000638906972807384,
+ "loss": 3.789350748062134,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.009501751512289047,
+ "learning_rate": 0.0006378310741758841,
+ "loss": 3.791079521179199,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.00936005637049675,
+ "learning_rate": 0.0006367558822635453,
+ "loss": 3.78501033782959,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.008370982483029366,
+ "learning_rate": 0.0006356813979366284,
+ "loss": 3.792647361755371,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.00782831758260727,
+ "learning_rate": 0.000634607622060827,
+ "loss": 3.772569179534912,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.00830658059567213,
+ "learning_rate": 0.0006335345555012627,
+ "loss": 3.7787656784057617,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.009015833027660847,
+ "learning_rate": 0.000632462199122486,
+ "loss": 3.787418842315674,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.008993860334157944,
+ "learning_rate": 0.0006313905537884752,
+ "loss": 3.818718433380127,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.007954884320497513,
+ "learning_rate": 0.0006303196203626344,
+ "loss": 3.8100008964538574,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.007742907386273146,
+ "learning_rate": 0.0006292493997077973,
+ "loss": 3.7786598205566406,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.008748230524361134,
+ "learning_rate": 0.0006281798926862195,
+ "loss": 3.798635482788086,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.008306438103318214,
+ "learning_rate": 0.000627111100159585,
+ "loss": 3.827213764190674,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.008027350530028343,
+ "learning_rate": 0.0006260430229889987,
+ "loss": 3.7682337760925293,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.007779776118695736,
+ "learning_rate": 0.0006249756620349931,
+ "loss": 3.7515475749969482,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.008748390711843967,
+ "learning_rate": 0.000623909018157521,
+ "loss": 3.7963085174560547,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.008587505668401718,
+ "learning_rate": 0.0006228430922159577,
+ "loss": 3.7767529487609863,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.007172323763370514,
+ "learning_rate": 0.0006217778850691007,
+ "loss": 3.79768705368042,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.007224083412438631,
+ "learning_rate": 0.000620713397575168,
+ "loss": 3.785390853881836,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.008209837600588799,
+ "learning_rate": 0.0006196496305917987,
+ "loss": 3.809828281402588,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.007760563865303993,
+ "learning_rate": 0.0006185865849760491,
+ "loss": 3.838716506958008,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.007362260017544031,
+ "learning_rate": 0.0006175242615843977,
+ "loss": 3.7840137481689453,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.008719335310161114,
+ "learning_rate": 0.0006164626612727381,
+ "loss": 3.8228893280029297,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.00856029987335205,
+ "learning_rate": 0.0006154017848963826,
+ "loss": 3.80448317527771,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.008078506216406822,
+ "learning_rate": 0.0006143416333100606,
+ "loss": 3.7881453037261963,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.007957885973155499,
+ "learning_rate": 0.0006132822073679164,
+ "loss": 3.814870834350586,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.009362339973449707,
+ "learning_rate": 0.0006122235079235114,
+ "loss": 3.798335075378418,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.008272632025182247,
+ "learning_rate": 0.0006111655358298194,
+ "loss": 3.8212976455688477,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.007548908703029156,
+ "learning_rate": 0.0006101082919392301,
+ "loss": 3.8312830924987793,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.007543765474110842,
+ "learning_rate": 0.0006090517771035457,
+ "loss": 3.8302183151245117,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.007520351093262434,
+ "learning_rate": 0.000607995992173981,
+ "loss": 3.7898802757263184,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.007168850861489773,
+ "learning_rate": 0.000606940938001163,
+ "loss": 3.8078577518463135,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.007420001085847616,
+ "learning_rate": 0.0006058866154351296,
+ "loss": 3.799571990966797,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.007621012162417173,
+ "learning_rate": 0.0006048330253253303,
+ "loss": 3.833843946456909,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.007661828305572271,
+ "learning_rate": 0.0006037801685206222,
+ "loss": 3.820556640625,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.007657836657017469,
+ "learning_rate": 0.0006027280458692738,
+ "loss": 3.798994541168213,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.007722760550677776,
+ "learning_rate": 0.0006016766582189613,
+ "loss": 3.8589251041412354,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.00789247453212738,
+ "learning_rate": 0.0006006260064167685,
+ "loss": 3.8673787117004395,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.009137817658483982,
+ "learning_rate": 0.0005995760913091874,
+ "loss": 3.801938056945801,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.008374894969165325,
+ "learning_rate": 0.0005985269137421141,
+ "loss": 3.820045232772827,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.3730878829956055,
+ "eval_runtime": 48.1607,
+ "eval_samples_per_second": 50.705,
+ "eval_steps_per_second": 1.599,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.009492472745478153,
+ "learning_rate": 0.0005974784745608536,
+ "loss": 3.738617420196533,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.010017318651080132,
+ "learning_rate": 0.0005964307746101135,
+ "loss": 3.7551231384277344,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.008958659134805202,
+ "learning_rate": 0.000595383814734007,
+ "loss": 3.7739763259887695,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.008034365251660347,
+ "learning_rate": 0.0005943375957760505,
+ "loss": 3.7678627967834473,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.008489780128002167,
+ "learning_rate": 0.000593292118579164,
+ "loss": 3.720778465270996,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.009107839316129684,
+ "learning_rate": 0.0005922473839856699,
+ "loss": 3.7224440574645996,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.009295578114688396,
+ "learning_rate": 0.0005912033928372906,
+ "loss": 3.7695257663726807,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.007896862924098969,
+ "learning_rate": 0.0005901601459751532,
+ "loss": 3.742567300796509,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.008227395825088024,
+ "learning_rate": 0.0005891176442397808,
+ "loss": 3.7547202110290527,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.009326796047389507,
+ "learning_rate": 0.0005880758884710992,
+ "loss": 3.724808931350708,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.007786506786942482,
+ "learning_rate": 0.0005870348795084319,
+ "loss": 3.7403640747070312,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.00821041688323021,
+ "learning_rate": 0.0005859946181905009,
+ "loss": 3.770012617111206,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.009544595144689083,
+ "learning_rate": 0.000584955105355427,
+ "loss": 3.7378525733947754,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.008973932825028896,
+ "learning_rate": 0.0005839163418407254,
+ "loss": 3.7184295654296875,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.009023688733577728,
+ "learning_rate": 0.00058287832848331,
+ "loss": 3.7388229370117188,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.008151066489517689,
+ "learning_rate": 0.0005818410661194889,
+ "loss": 3.7632875442504883,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.008495030924677849,
+ "learning_rate": 0.0005808045555849668,
+ "loss": 3.8114736080169678,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.008739210665225983,
+ "learning_rate": 0.0005797687977148394,
+ "loss": 3.7647829055786133,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.008062886074185371,
+ "learning_rate": 0.0005787337933435999,
+ "loss": 3.782172441482544,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.007615398149937391,
+ "learning_rate": 0.0005776995433051326,
+ "loss": 3.7564642429351807,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.007938290014863014,
+ "learning_rate": 0.0005766660484327133,
+ "loss": 3.756208896636963,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.008517666719853878,
+ "learning_rate": 0.0005756333095590102,
+ "loss": 3.76729416847229,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.007426336407661438,
+ "learning_rate": 0.0005746013275160827,
+ "loss": 3.74001407623291,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.007664753124117851,
+ "learning_rate": 0.0005735701031353804,
+ "loss": 3.7819876670837402,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.007903067395091057,
+ "learning_rate": 0.0005725396372477414,
+ "loss": 3.781219959259033,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.007251975126564503,
+ "learning_rate": 0.0005715099306833931,
+ "loss": 3.763620615005493,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.007422176189720631,
+ "learning_rate": 0.0005704809842719533,
+ "loss": 3.775099754333496,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.009161600843071938,
+ "learning_rate": 0.0005694527988424239,
+ "loss": 3.792268991470337,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.008330529555678368,
+ "learning_rate": 0.0005684253752231958,
+ "loss": 3.7967824935913086,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.0073860278353095055,
+ "learning_rate": 0.0005673987142420458,
+ "loss": 3.76505970954895,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.008462096564471722,
+ "learning_rate": 0.0005663728167261363,
+ "loss": 3.740450143814087,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.00838128849864006,
+ "learning_rate": 0.0005653476835020138,
+ "loss": 3.7863235473632812,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.007381014991551638,
+ "learning_rate": 0.0005643233153956092,
+ "loss": 3.772179126739502,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.007363761775195599,
+ "learning_rate": 0.0005632997132322392,
+ "loss": 3.7793867588043213,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.008684969507157803,
+ "learning_rate": 0.0005622768778366001,
+ "loss": 3.7736339569091797,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.007934059016406536,
+ "learning_rate": 0.0005612548100327722,
+ "loss": 3.8040053844451904,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.007136524189263582,
+ "learning_rate": 0.0005602335106442172,
+ "loss": 3.7589468955993652,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.00789916142821312,
+ "learning_rate": 0.0005592129804937779,
+ "loss": 3.7953195571899414,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.007919179275631905,
+ "learning_rate": 0.0005581932204036761,
+ "loss": 3.7539429664611816,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.007734223268926144,
+ "learning_rate": 0.0005571742311955147,
+ "loss": 3.7620325088500977,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.007005508057773113,
+ "learning_rate": 0.0005561560136902749,
+ "loss": 3.762538433074951,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.008137146942317486,
+ "learning_rate": 0.0005551385687083161,
+ "loss": 3.773789405822754,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.007310070563107729,
+ "learning_rate": 0.0005541218970693757,
+ "loss": 3.8077168464660645,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.007335536181926727,
+ "learning_rate": 0.0005531059995925661,
+ "loss": 3.7761921882629395,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.008022893220186234,
+ "learning_rate": 0.00055209087709638,
+ "loss": 3.7769880294799805,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.008027128875255585,
+ "learning_rate": 0.0005510765303986814,
+ "loss": 3.7709951400756836,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.00697955721989274,
+ "learning_rate": 0.0005500629603167118,
+ "loss": 3.7855210304260254,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.007455077022314072,
+ "learning_rate": 0.0005490501676670861,
+ "loss": 3.7850632667541504,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.0069657545536756516,
+ "learning_rate": 0.0005480381532657932,
+ "loss": 3.742671251296997,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.007609451189637184,
+ "learning_rate": 0.0005470269179281953,
+ "loss": 3.793771266937256,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.007437967229634523,
+ "learning_rate": 0.0005460164624690255,
+ "loss": 3.778846263885498,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.007123028859496117,
+ "learning_rate": 0.00054500678770239,
+ "loss": 3.768375873565674,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.008176218718290329,
+ "learning_rate": 0.0005439978944417654,
+ "loss": 3.82069730758667,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.006943001411855221,
+ "learning_rate": 0.000542989783499999,
+ "loss": 3.7731552124023438,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.0072205523028969765,
+ "learning_rate": 0.0005419824556893069,
+ "loss": 3.7724058628082275,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.007656400091946125,
+ "learning_rate": 0.0005409759118212758,
+ "loss": 3.780989646911621,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.006972154602408409,
+ "learning_rate": 0.0005399701527068599,
+ "loss": 3.7971787452697754,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.007674331311136484,
+ "learning_rate": 0.0005389651791563806,
+ "loss": 3.798673391342163,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.006952045019716024,
+ "learning_rate": 0.0005379609919795268,
+ "loss": 3.7982821464538574,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.007882075384259224,
+ "learning_rate": 0.0005369575919853545,
+ "loss": 3.825962781906128,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.007143386639654636,
+ "learning_rate": 0.0005359549799822854,
+ "loss": 3.826598644256592,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.007587207015603781,
+ "learning_rate": 0.0005349531567781041,
+ "loss": 3.80623197555542,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.007677781395614147,
+ "learning_rate": 0.0005339521231799632,
+ "loss": 3.7987256050109863,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.007372343447059393,
+ "learning_rate": 0.0005329518799943773,
+ "loss": 3.7733969688415527,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.00861304346472025,
+ "learning_rate": 0.0005319524280272231,
+ "loss": 3.7903685569763184,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.00816237274557352,
+ "learning_rate": 0.0005309537680837418,
+ "loss": 3.77058744430542,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.007624274119734764,
+ "learning_rate": 0.0005299559009685356,
+ "loss": 3.787982940673828,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.007416521664708853,
+ "learning_rate": 0.0005289588274855683,
+ "loss": 3.7991762161254883,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.007461614441126585,
+ "learning_rate": 0.0005279625484381632,
+ "loss": 3.8036863803863525,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.007766335736960173,
+ "learning_rate": 0.0005269670646290043,
+ "loss": 3.809243679046631,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.0073394919745624065,
+ "learning_rate": 0.0005259723768601365,
+ "loss": 3.756606101989746,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.008771199733018875,
+ "learning_rate": 0.0005249784859329602,
+ "loss": 3.773951768875122,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.370637893676758,
+ "eval_runtime": 49.4904,
+ "eval_samples_per_second": 49.343,
+ "eval_steps_per_second": 1.556,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.008164486847817898,
+ "learning_rate": 0.0005239853926482358,
+ "loss": 3.7619595527648926,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.007980132475495338,
+ "learning_rate": 0.0005229930978060806,
+ "loss": 3.7398204803466797,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.0077056242153048515,
+ "learning_rate": 0.0005220016022059692,
+ "loss": 3.699580669403076,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.0074698119424283504,
+ "learning_rate": 0.0005210109066467308,
+ "loss": 3.7299928665161133,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.008384468965232372,
+ "learning_rate": 0.0005200210119265512,
+ "loss": 3.704989433288574,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.008066141977906227,
+ "learning_rate": 0.0005190319188429709,
+ "loss": 3.751697540283203,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.007213297300040722,
+ "learning_rate": 0.0005180436281928841,
+ "loss": 3.740692377090454,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.007343464996665716,
+ "learning_rate": 0.0005170561407725396,
+ "loss": 3.7060189247131348,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.00783725269138813,
+ "learning_rate": 0.0005160694573775363,
+ "loss": 3.7414426803588867,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.007640232797712088,
+ "learning_rate": 0.0005150835788028294,
+ "loss": 3.7451486587524414,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.008129071444272995,
+ "learning_rate": 0.0005140985058427218,
+ "loss": 3.734839916229248,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.006920623127371073,
+ "learning_rate": 0.0005131142392908697,
+ "loss": 3.7138614654541016,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.00798740889877081,
+ "learning_rate": 0.0005121307799402787,
+ "loss": 3.76496958732605,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.006689135450869799,
+ "learning_rate": 0.0005111481285833045,
+ "loss": 3.7707948684692383,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.007544311694800854,
+ "learning_rate": 0.0005101662860116516,
+ "loss": 3.7393479347229004,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.007487910334020853,
+ "learning_rate": 0.0005091852530163717,
+ "loss": 3.7486305236816406,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.00772349676117301,
+ "learning_rate": 0.0005082050303878673,
+ "loss": 3.753629207611084,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.007883663289248943,
+ "learning_rate": 0.0005072256189158844,
+ "loss": 3.7497611045837402,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.008153906092047691,
+ "learning_rate": 0.0005062470193895178,
+ "loss": 3.7459521293640137,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.007083641365170479,
+ "learning_rate": 0.0005052692325972075,
+ "loss": 3.725599765777588,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.007102683652192354,
+ "learning_rate": 0.0005042922593267385,
+ "loss": 3.7724428176879883,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.008104706183075905,
+ "learning_rate": 0.0005033161003652409,
+ "loss": 3.7386724948883057,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.0073765418492257595,
+ "learning_rate": 0.0005023407564991878,
+ "loss": 3.7478132247924805,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.007110628765076399,
+ "learning_rate": 0.0005013662285143963,
+ "loss": 3.771925687789917,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.007323848083615303,
+ "learning_rate": 0.000500392517196026,
+ "loss": 3.7460904121398926,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.007830807007849216,
+ "learning_rate": 0.0004994196233285786,
+ "loss": 3.751222610473633,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.007710007485002279,
+ "learning_rate": 0.0004984475476958973,
+ "loss": 3.739858627319336,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.007293973118066788,
+ "learning_rate": 0.0004974762910811655,
+ "loss": 3.7410120964050293,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.0073824538849294186,
+ "learning_rate": 0.0004965058542669078,
+ "loss": 3.711397647857666,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.007156423293054104,
+ "learning_rate": 0.0004955362380349866,
+ "loss": 3.7881593704223633,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.007705818396061659,
+ "learning_rate": 0.0004945674431666047,
+ "loss": 3.7639269828796387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.0075300210155546665,
+ "learning_rate": 0.0004935994704423026,
+ "loss": 3.7562999725341797,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.007137116510421038,
+ "learning_rate": 0.000492632320641959,
+ "loss": 3.7659149169921875,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.007225630804896355,
+ "learning_rate": 0.0004916659945447878,
+ "loss": 3.7625155448913574,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.008027992211282253,
+ "learning_rate": 0.0004907004929293408,
+ "loss": 3.746478796005249,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.007582477759569883,
+ "learning_rate": 0.0004897358165735066,
+ "loss": 3.759408950805664,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.007575645577162504,
+ "learning_rate": 0.000488771966254506,
+ "loss": 3.7430644035339355,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.008404700085520744,
+ "learning_rate": 0.0004878089427488963,
+ "loss": 3.7727198600769043,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.007118664216250181,
+ "learning_rate": 0.00048684674683256815,
+ "loss": 3.747460126876831,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.007220778148621321,
+ "learning_rate": 0.00048588537928074584,
+ "loss": 3.746413230895996,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.0076194945722818375,
+ "learning_rate": 0.0004849248408679849,
+ "loss": 3.7808125019073486,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.007285538129508495,
+ "learning_rate": 0.00048396513236817387,
+ "loss": 3.7750301361083984,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.007698678877204657,
+ "learning_rate": 0.00048300625455453383,
+ "loss": 3.7474594116210938,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.007687424309551716,
+ "learning_rate": 0.0004820482081996141,
+ "loss": 3.7929720878601074,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.007363605313003063,
+ "learning_rate": 0.0004810909940752959,
+ "loss": 3.75911021232605,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.007791243493556976,
+ "learning_rate": 0.00048013461295278836,
+ "loss": 3.7567291259765625,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.007375436834990978,
+ "learning_rate": 0.0004791790656026323,
+ "loss": 3.7784414291381836,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.007330593653023243,
+ "learning_rate": 0.0004782243527946934,
+ "loss": 3.7533810138702393,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.006926248781383038,
+ "learning_rate": 0.00047727047529816694,
+ "loss": 3.7795910835266113,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.007102136500179768,
+ "learning_rate": 0.00047631743388157496,
+ "loss": 3.812549114227295,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.006676106248050928,
+ "learning_rate": 0.00047536522931276555,
+ "loss": 3.7578792572021484,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.006926767062395811,
+ "learning_rate": 0.0004744138623589134,
+ "loss": 3.782416820526123,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.006894777063280344,
+ "learning_rate": 0.000473463333786516,
+ "loss": 3.7840042114257812,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.007393381092697382,
+ "learning_rate": 0.00047251364436139923,
+ "loss": 3.7532780170440674,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.006510896608233452,
+ "learning_rate": 0.0004715647948487092,
+ "loss": 3.7618460655212402,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.007283828221261501,
+ "learning_rate": 0.0004706167860129171,
+ "loss": 3.768498420715332,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.00655644154176116,
+ "learning_rate": 0.0004696696186178168,
+ "loss": 3.7724053859710693,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.007338619790971279,
+ "learning_rate": 0.0004687232934265239,
+ "loss": 3.7814595699310303,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.006839970126748085,
+ "learning_rate": 0.0004677778112014758,
+ "loss": 3.771808624267578,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.006776005961000919,
+ "learning_rate": 0.0004668331727044297,
+ "loss": 3.786924123764038,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.007011357229202986,
+ "learning_rate": 0.00046588937869646415,
+ "loss": 3.787672996520996,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.006482086144387722,
+ "learning_rate": 0.0004649464299379771,
+ "loss": 3.7794556617736816,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.0069654518738389015,
+ "learning_rate": 0.000464004327188685,
+ "loss": 3.7520360946655273,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.006526473443955183,
+ "learning_rate": 0.0004630630712076235,
+ "loss": 3.734144926071167,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.006822322495281696,
+ "learning_rate": 0.00046212266275314504,
+ "loss": 3.7614707946777344,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.006827171426266432,
+ "learning_rate": 0.00046118310258292045,
+ "loss": 3.755218029022217,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.00706718722358346,
+ "learning_rate": 0.00046024439145393546,
+ "loss": 3.7397823333740234,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.00689381267875433,
+ "learning_rate": 0.00045930653012249314,
+ "loss": 3.758922576904297,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.007069882471114397,
+ "learning_rate": 0.0004583695193442113,
+ "loss": 3.759767532348633,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.006889182608574629,
+ "learning_rate": 0.0004574333598740228,
+ "loss": 3.7555460929870605,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.006866258103400469,
+ "learning_rate": 0.00045649805246617475,
+ "loss": 3.762068271636963,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.008696220815181732,
+ "learning_rate": 0.000455563597874226,
+ "loss": 3.7110466957092285,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.369833469390869,
+ "eval_runtime": 48.7852,
+ "eval_samples_per_second": 50.056,
+ "eval_steps_per_second": 1.578,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.0077184769324958324,
+ "learning_rate": 0.0004546299968510516,
+ "loss": 3.704669713973999,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.00846019759774208,
+ "learning_rate": 0.0004536972501488358,
+ "loss": 3.701132297515869,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.008113945834338665,
+ "learning_rate": 0.00045276535851907613,
+ "loss": 3.699582099914551,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.0072370911948382854,
+ "learning_rate": 0.0004518343227125807,
+ "loss": 3.698126792907715,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.00817187875509262,
+ "learning_rate": 0.0004509041434794685,
+ "loss": 3.738393783569336,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.007495569530874491,
+ "learning_rate": 0.00044997482156916845,
+ "loss": 3.727083683013916,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.007855897769331932,
+ "learning_rate": 0.0004490463577304171,
+ "loss": 3.716137170791626,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.007817867211997509,
+ "learning_rate": 0.0004481187527112628,
+ "loss": 3.750791072845459,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.007712380029261112,
+ "learning_rate": 0.00044719200725905887,
+ "loss": 3.720554828643799,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.00859921332448721,
+ "learning_rate": 0.00044626612212046816,
+ "loss": 3.742496967315674,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.007692081853747368,
+ "learning_rate": 0.0004453410980414582,
+ "loss": 3.703803300857544,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.007734883576631546,
+ "learning_rate": 0.00044441693576730565,
+ "loss": 3.7371301651000977,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.009022004902362823,
+ "learning_rate": 0.0004434936360425916,
+ "loss": 3.727294683456421,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.008058780804276466,
+ "learning_rate": 0.00044257119961120067,
+ "loss": 3.7583885192871094,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.00698013324290514,
+ "learning_rate": 0.00044164962721632376,
+ "loss": 3.7312440872192383,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.007999536581337452,
+ "learning_rate": 0.00044072891960045523,
+ "loss": 3.6972708702087402,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.008125107735395432,
+ "learning_rate": 0.0004398090775053929,
+ "loss": 3.7123398780822754,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.006736223120242357,
+ "learning_rate": 0.0004388901016722357,
+ "loss": 3.749678134918213,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.009154478088021278,
+ "learning_rate": 0.000437971992841387,
+ "loss": 3.722728967666626,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.008070871233940125,
+ "learning_rate": 0.00043705475175255087,
+ "loss": 3.7297017574310303,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.007498566992580891,
+ "learning_rate": 0.0004361383791447307,
+ "loss": 3.7286131381988525,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.007642887998372316,
+ "learning_rate": 0.00043522287575623183,
+ "loss": 3.750898838043213,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.007800443097949028,
+ "learning_rate": 0.0004343082423246588,
+ "loss": 3.719512462615967,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.007393732666969299,
+ "learning_rate": 0.0004333944795869159,
+ "loss": 3.7136006355285645,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.008559555746614933,
+ "learning_rate": 0.0004324815882792042,
+ "loss": 3.7087020874023438,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.007379850372672081,
+ "learning_rate": 0.0004315695691370241,
+ "loss": 3.7638165950775146,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.007374506443738937,
+ "learning_rate": 0.00043065842289517303,
+ "loss": 3.7208821773529053,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.00755207147449255,
+ "learning_rate": 0.0004297481502877451,
+ "loss": 3.738490581512451,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.007752610836178064,
+ "learning_rate": 0.0004288387520481306,
+ "loss": 3.7401833534240723,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.006907653529196978,
+ "learning_rate": 0.0004279302289090151,
+ "loss": 3.7754998207092285,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.007771312724798918,
+ "learning_rate": 0.0004270225816023797,
+ "loss": 3.73626446723938,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.007395982276648283,
+ "learning_rate": 0.0004261158108594987,
+ "loss": 3.7159175872802734,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.007196040358394384,
+ "learning_rate": 0.0004252099174109411,
+ "loss": 3.750246047973633,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.008241173811256886,
+ "learning_rate": 0.0004243049019865692,
+ "loss": 3.70859694480896,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.0072947596199810505,
+ "learning_rate": 0.00042340076531553755,
+ "loss": 3.7385268211364746,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.00782464537769556,
+ "learning_rate": 0.00042249750812629297,
+ "loss": 3.75885272026062,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.007813232019543648,
+ "learning_rate": 0.0004215951311465722,
+ "loss": 3.71810245513916,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.006900500506162643,
+ "learning_rate": 0.00042069363510340604,
+ "loss": 3.7365174293518066,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.007812316995114088,
+ "learning_rate": 0.0004197930207231119,
+ "loss": 3.774794101715088,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.007018622010946274,
+ "learning_rate": 0.00041889328873129933,
+ "loss": 3.7112040519714355,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.006895892322063446,
+ "learning_rate": 0.0004179944398528659,
+ "loss": 3.7497243881225586,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.0071940235793590546,
+ "learning_rate": 0.0004170964748119979,
+ "loss": 3.7566123008728027,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.006776201073080301,
+ "learning_rate": 0.0004161993943321702,
+ "loss": 3.7683968544006348,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.006886148825287819,
+ "learning_rate": 0.0004153031991361428,
+ "loss": 3.759763240814209,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.006466342136263847,
+ "learning_rate": 0.0004144078899459659,
+ "loss": 3.736006498336792,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.007055724039673805,
+ "learning_rate": 0.0004135134674829725,
+ "loss": 3.70645809173584,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.006948428228497505,
+ "learning_rate": 0.00041261993246778303,
+ "loss": 3.7365002632141113,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.006491438951343298,
+ "learning_rate": 0.00041172728562030255,
+ "loss": 3.735001564025879,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.006851755082607269,
+ "learning_rate": 0.0004108355276597209,
+ "loss": 3.720893383026123,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.007266475819051266,
+ "learning_rate": 0.00040994465930451114,
+ "loss": 3.7746100425720215,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.007112737279385328,
+ "learning_rate": 0.00040905468127242926,
+ "loss": 3.7441720962524414,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.006514255423098803,
+ "learning_rate": 0.000408165594280515,
+ "loss": 3.7333898544311523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.007172788958996534,
+ "learning_rate": 0.0004072773990450896,
+ "loss": 3.7675576210021973,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.00684684282168746,
+ "learning_rate": 0.0004063900962817564,
+ "loss": 3.7141308784484863,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.006578145083039999,
+ "learning_rate": 0.00040550368670539777,
+ "loss": 3.7256317138671875,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.007047492545098066,
+ "learning_rate": 0.00040461817103017926,
+ "loss": 3.7618556022644043,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.007036720868200064,
+ "learning_rate": 0.00040373354996954487,
+ "loss": 3.7397944927215576,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.00745626725256443,
+ "learning_rate": 0.00040284982423621714,
+ "loss": 3.7831153869628906,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.007454817183315754,
+ "learning_rate": 0.00040196699454219797,
+ "loss": 3.7409911155700684,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.0075650583021342754,
+ "learning_rate": 0.00040108506159876756,
+ "loss": 3.7074005603790283,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.006908242590725422,
+ "learning_rate": 0.0004002040261164838,
+ "loss": 3.767744779586792,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.0072857895866036415,
+ "learning_rate": 0.0003993238888051801,
+ "loss": 3.7349488735198975,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.0072942771948874,
+ "learning_rate": 0.00039844465037396716,
+ "loss": 3.75968861579895,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.0071496241725981236,
+ "learning_rate": 0.00039756631153123296,
+ "loss": 3.8237380981445312,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.006803249940276146,
+ "learning_rate": 0.00039668887298463783,
+ "loss": 3.7468650341033936,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.007433352991938591,
+ "learning_rate": 0.00039581233544111867,
+ "loss": 3.783236503601074,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.006633558310568333,
+ "learning_rate": 0.0003949366996068858,
+ "loss": 3.75020694732666,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.007415154483169317,
+ "learning_rate": 0.00039406196618742376,
+ "loss": 3.7545979022979736,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.006997843272984028,
+ "learning_rate": 0.0003931881358874886,
+ "loss": 3.7404396533966064,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.007417330518364906,
+ "learning_rate": 0.0003923152094111095,
+ "loss": 3.7575364112854004,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.006936701480299234,
+ "learning_rate": 0.0003914431874615889,
+ "loss": 3.771845817565918,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.008293208666145802,
+ "learning_rate": 0.00039057207074149795,
+ "loss": 3.746669292449951,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.370476245880127,
+ "eval_runtime": 50.9368,
+ "eval_samples_per_second": 47.942,
+ "eval_steps_per_second": 1.512,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.0077544888481497765,
+ "learning_rate": 0.0003897018599526799,
+ "loss": 3.6915597915649414,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.008674043230712414,
+ "learning_rate": 0.000388832555796248,
+ "loss": 3.67698073387146,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.00724575063213706,
+ "learning_rate": 0.00038796415897258523,
+ "loss": 3.6838502883911133,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.00849415734410286,
+ "learning_rate": 0.00038709667018134246,
+ "loss": 3.69064998626709,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.00769044877961278,
+ "learning_rate": 0.0003862300901214403,
+ "loss": 3.670502185821533,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.007846124470233917,
+ "learning_rate": 0.00038536441949106647,
+ "loss": 3.767771005630493,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.007555077783763409,
+ "learning_rate": 0.000384499658987676,
+ "loss": 3.7001633644104004,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.007834251970052719,
+ "learning_rate": 0.0003836358093079915,
+ "loss": 3.704331874847412,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.007608097977936268,
+ "learning_rate": 0.00038277287114799974,
+ "loss": 3.7108519077301025,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.007420446258038282,
+ "learning_rate": 0.0003819108452029561,
+ "loss": 3.697843313217163,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.008091001771390438,
+ "learning_rate": 0.000381049732167378,
+ "loss": 3.7380661964416504,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.007265515625476837,
+ "learning_rate": 0.00038018953273504907,
+ "loss": 3.680748462677002,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.007735357619822025,
+ "learning_rate": 0.00037933024759901636,
+ "loss": 3.6648077964782715,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.008188895881175995,
+ "learning_rate": 0.0003784718774515907,
+ "loss": 3.706609010696411,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.00816027820110321,
+ "learning_rate": 0.00037761442298434554,
+ "loss": 3.7368297576904297,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.006931936834007502,
+ "learning_rate": 0.00037675788488811566,
+ "loss": 3.7121026515960693,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.007615855894982815,
+ "learning_rate": 0.00037590226385299897,
+ "loss": 3.7379372119903564,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.0073505607433617115,
+ "learning_rate": 0.0003750475605683535,
+ "loss": 3.7330894470214844,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.007151979021728039,
+ "learning_rate": 0.00037419377572279897,
+ "loss": 3.713761568069458,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.007310053799301386,
+ "learning_rate": 0.00037334091000421326,
+ "loss": 3.705451011657715,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.0072968751192092896,
+ "learning_rate": 0.0003724889640997357,
+ "loss": 3.7550907135009766,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.007197601720690727,
+ "learning_rate": 0.00037163793869576404,
+ "loss": 3.7346463203430176,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.007823911495506763,
+ "learning_rate": 0.00037078783447795307,
+ "loss": 3.7257699966430664,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.007118315435945988,
+ "learning_rate": 0.0003699386521312171,
+ "loss": 3.7014496326446533,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.0069230347871780396,
+ "learning_rate": 0.00036909039233972666,
+ "loss": 3.6709237098693848,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.007836844772100449,
+ "learning_rate": 0.00036824305578691,
+ "loss": 3.695725440979004,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.007123398594558239,
+ "learning_rate": 0.00036739664315544956,
+ "loss": 3.7278594970703125,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.007792143151164055,
+ "learning_rate": 0.00036655115512728586,
+ "loss": 3.7291927337646484,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.007245976477861404,
+ "learning_rate": 0.00036570659238361345,
+ "loss": 3.7183046340942383,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.007778054103255272,
+ "learning_rate": 0.0003648629556048804,
+ "loss": 3.7485761642456055,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.007539471611380577,
+ "learning_rate": 0.00036402024547079007,
+ "loss": 3.7273335456848145,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.007554007228463888,
+ "learning_rate": 0.0003631784626602987,
+ "loss": 3.684359073638916,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.007862667553126812,
+ "learning_rate": 0.00036233760785161606,
+ "loss": 3.7223072052001953,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.00869659148156643,
+ "learning_rate": 0.0003614976817222028,
+ "loss": 3.7475168704986572,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.00696196174249053,
+ "learning_rate": 0.00036065868494877226,
+ "loss": 3.7773914337158203,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.008952745236456394,
+ "learning_rate": 0.00035982061820729005,
+ "loss": 3.6761155128479004,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.008341601118445396,
+ "learning_rate": 0.00035898348217297034,
+ "loss": 3.746553421020508,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.007268146146088839,
+ "learning_rate": 0.00035814727752027885,
+ "loss": 3.6921701431274414,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.007401647046208382,
+ "learning_rate": 0.00035731200492293064,
+ "loss": 3.722433090209961,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.006969883106648922,
+ "learning_rate": 0.00035647766505389006,
+ "loss": 3.7449448108673096,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.007616426330059767,
+ "learning_rate": 0.00035564425858536895,
+ "loss": 3.725113868713379,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.006883042864501476,
+ "learning_rate": 0.00035481178618882814,
+ "loss": 3.717106342315674,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.007106961216777563,
+ "learning_rate": 0.00035398024853497574,
+ "loss": 3.74045467376709,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.007402827963232994,
+ "learning_rate": 0.00035314964629376664,
+ "loss": 3.7288095951080322,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.006770852487534285,
+ "learning_rate": 0.00035231998013440227,
+ "loss": 3.7663369178771973,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.007622968405485153,
+ "learning_rate": 0.0003514912507253287,
+ "loss": 3.747084140777588,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.006876434665173292,
+ "learning_rate": 0.00035066345873423975,
+ "loss": 3.7591025829315186,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.007071215659379959,
+ "learning_rate": 0.00034983660482807116,
+ "loss": 3.7583885192871094,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.007130554411560297,
+ "learning_rate": 0.00034901068967300465,
+ "loss": 3.6950902938842773,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.006981188431382179,
+ "learning_rate": 0.00034818571393446524,
+ "loss": 3.7390527725219727,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.0067793820053339005,
+ "learning_rate": 0.0003473616782771209,
+ "loss": 3.703673839569092,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.00674026133492589,
+ "learning_rate": 0.0003465385833648827,
+ "loss": 3.6988182067871094,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.006553220096975565,
+ "learning_rate": 0.0003457164298609021,
+ "loss": 3.7367048263549805,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.007079749833792448,
+ "learning_rate": 0.000344895218427575,
+ "loss": 3.773226022720337,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.006091540213674307,
+ "learning_rate": 0.00034407494972653547,
+ "loss": 3.747929096221924,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.006783034652471542,
+ "learning_rate": 0.00034325562441865935,
+ "loss": 3.7154481410980225,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.006683558225631714,
+ "learning_rate": 0.0003424372431640625,
+ "loss": 3.751605987548828,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.0066695502027869225,
+ "learning_rate": 0.0003416198066220997,
+ "loss": 3.7561278343200684,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.006601935718208551,
+ "learning_rate": 0.0003408033154513655,
+ "loss": 3.754878520965576,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.006911450996994972,
+ "learning_rate": 0.00033998777030969085,
+ "loss": 3.69533109664917,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.0070345234125852585,
+ "learning_rate": 0.0003391731718541465,
+ "loss": 3.7328338623046875,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.007238153368234634,
+ "learning_rate": 0.00033835952074103964,
+ "loss": 3.7370123863220215,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.007295276038348675,
+ "learning_rate": 0.0003375468176259142,
+ "loss": 3.7123947143554688,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.006471444386988878,
+ "learning_rate": 0.0003367350631635506,
+ "loss": 3.73347806930542,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.007136075291782618,
+ "learning_rate": 0.00033592425800796447,
+ "loss": 3.7337799072265625,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.0065771411173045635,
+ "learning_rate": 0.00033511440281240725,
+ "loss": 3.733506441116333,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.006566311698406935,
+ "learning_rate": 0.0003343054982293643,
+ "loss": 3.7325632572174072,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.006389300804585218,
+ "learning_rate": 0.00033349754491055535,
+ "loss": 3.7027206420898438,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.006425539031624794,
+ "learning_rate": 0.00033269054350693386,
+ "loss": 3.751466989517212,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.006735894829034805,
+ "learning_rate": 0.0003318844946686865,
+ "loss": 3.731363296508789,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.0061791036278009415,
+ "learning_rate": 0.00033107939904523264,
+ "loss": 3.7592201232910156,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.007923831231892109,
+ "learning_rate": 0.0003302752572852219,
+ "loss": 3.7247159481048584,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.371238708496094,
+ "eval_runtime": 51.312,
+ "eval_samples_per_second": 47.591,
+ "eval_steps_per_second": 1.501,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.007161229848861694,
+ "learning_rate": 0.0003294720700365385,
+ "loss": 3.6985998153686523,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.007182417903095484,
+ "learning_rate": 0.00032866983794629457,
+ "loss": 3.7116494178771973,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.0072847092524170876,
+ "learning_rate": 0.00032786856166083463,
+ "loss": 3.6759095191955566,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.006694195792078972,
+ "learning_rate": 0.0003270682418257323,
+ "loss": 3.6635122299194336,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.0073003629222512245,
+ "learning_rate": 0.0003262688790857914,
+ "loss": 3.7130517959594727,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.006921815685927868,
+ "learning_rate": 0.0003254704740850432,
+ "loss": 3.6923139095306396,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.007585546933114529,
+ "learning_rate": 0.00032467302746674805,
+ "loss": 3.680659770965576,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.0069264741614460945,
+ "learning_rate": 0.0003238765398733956,
+ "loss": 3.7042646408081055,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.007581264711916447,
+ "learning_rate": 0.0003230810119467003,
+ "loss": 3.6890318393707275,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.007325596176087856,
+ "learning_rate": 0.00032228644432760544,
+ "loss": 3.728785753250122,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.006920192390680313,
+ "learning_rate": 0.0003214928376562787,
+ "loss": 3.765744686126709,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.007247452158480883,
+ "learning_rate": 0.00032070019257211596,
+ "loss": 3.699038505554199,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.0073222690261900425,
+ "learning_rate": 0.0003199085097137361,
+ "loss": 3.6657447814941406,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.0074598658829927444,
+ "learning_rate": 0.000319117789718984,
+ "loss": 3.6929068565368652,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.007144949398934841,
+ "learning_rate": 0.0003183280332249286,
+ "loss": 3.7166645526885986,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.0068069938570261,
+ "learning_rate": 0.0003175392408678621,
+ "loss": 3.650606155395508,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.007320682052522898,
+ "learning_rate": 0.0003167514132833012,
+ "loss": 3.69647216796875,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.007024327293038368,
+ "learning_rate": 0.00031596455110598264,
+ "loss": 3.6764140129089355,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.007796750869601965,
+ "learning_rate": 0.0003151786549698689,
+ "loss": 3.6841444969177246,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.007526555098593235,
+ "learning_rate": 0.0003143937255081411,
+ "loss": 3.7181622982025146,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.00689823878929019,
+ "learning_rate": 0.0003136097633532032,
+ "loss": 3.6805999279022217,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.007449387572705746,
+ "learning_rate": 0.00031282676913667903,
+ "loss": 3.719672679901123,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.0072627803310751915,
+ "learning_rate": 0.00031204474348941336,
+ "loss": 3.692244291305542,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.0073951175436377525,
+ "learning_rate": 0.00031126368704147026,
+ "loss": 3.734830379486084,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 0.007486097048968077,
+ "learning_rate": 0.0003104836004221318,
+ "loss": 3.6959404945373535,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.007235408294945955,
+ "learning_rate": 0.0003097044842599005,
+ "loss": 3.714496612548828,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.0074761975556612015,
+ "learning_rate": 0.00030892633918249584,
+ "loss": 3.6801066398620605,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.007354153320193291,
+ "learning_rate": 0.0003081491658168555,
+ "loss": 3.7053956985473633,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 0.0077071161940693855,
+ "learning_rate": 0.0003073729647891338,
+ "loss": 3.724217653274536,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.007286814507097006,
+ "learning_rate": 0.00030659773672470227,
+ "loss": 3.717681646347046,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.00741427019238472,
+ "learning_rate": 0.00030582348224814803,
+ "loss": 3.7178781032562256,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.007106068544089794,
+ "learning_rate": 0.00030505020198327355,
+ "loss": 3.671164035797119,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.006869092117995024,
+ "learning_rate": 0.0003042778965530969,
+ "loss": 3.7310664653778076,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.00765156140550971,
+ "learning_rate": 0.0003035065665798508,
+ "loss": 3.7417078018188477,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.006932055577635765,
+ "learning_rate": 0.00030273621268498183,
+ "loss": 3.6812591552734375,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.008092030882835388,
+ "learning_rate": 0.00030196683548915055,
+ "loss": 3.69986891746521,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.00719967158511281,
+ "learning_rate": 0.00030119843561222915,
+ "loss": 3.7247209548950195,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.007408039178699255,
+ "learning_rate": 0.00030043101367330504,
+ "loss": 3.7106001377105713,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.007386322598904371,
+ "learning_rate": 0.00029966457029067503,
+ "loss": 3.7073440551757812,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.007485459093004465,
+ "learning_rate": 0.00029889910608184924,
+ "loss": 3.730858325958252,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.007635290268808603,
+ "learning_rate": 0.00029813462166354814,
+ "loss": 3.6909053325653076,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0069068302400410175,
+ "learning_rate": 0.0002973711176517032,
+ "loss": 3.7286834716796875,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.00698851840570569,
+ "learning_rate": 0.0002966085946614562,
+ "loss": 3.6587166786193848,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.007312684785574675,
+ "learning_rate": 0.0002958470533071569,
+ "loss": 3.716031789779663,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.006896916776895523,
+ "learning_rate": 0.00029508649420236687,
+ "loss": 3.7105724811553955,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.007188099902123213,
+ "learning_rate": 0.00029432691795985394,
+ "loss": 3.712202548980713,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.006414467468857765,
+ "learning_rate": 0.000293568325191595,
+ "loss": 3.71462345123291,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.006734516471624374,
+ "learning_rate": 0.00029281071650877496,
+ "loss": 3.7277159690856934,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.0067634680308401585,
+ "learning_rate": 0.00029205409252178524,
+ "loss": 3.6998372077941895,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.006864939816296101,
+ "learning_rate": 0.00029129845384022446,
+ "loss": 3.7656338214874268,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.006820365786552429,
+ "learning_rate": 0.00029054380107289626,
+ "loss": 3.7018871307373047,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.006954800803214312,
+ "learning_rate": 0.00028979013482781114,
+ "loss": 3.7410342693328857,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.0072957612574100494,
+ "learning_rate": 0.00028903745571218453,
+ "loss": 3.6964735984802246,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.006805056240409613,
+ "learning_rate": 0.00028828576433243675,
+ "loss": 3.6988587379455566,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.007712861057370901,
+ "learning_rate": 0.0002875350612941908,
+ "loss": 3.73240327835083,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.006683751940727234,
+ "learning_rate": 0.0002867853472022755,
+ "loss": 3.6936707496643066,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.007726171053946018,
+ "learning_rate": 0.0002860366226607223,
+ "loss": 3.7266287803649902,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.007187390234321356,
+ "learning_rate": 0.00028528888827276396,
+ "loss": 3.7381067276000977,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.007067828439176083,
+ "learning_rate": 0.00028454214464083706,
+ "loss": 3.71884822845459,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.007044029422104359,
+ "learning_rate": 0.00028379639236657966,
+ "loss": 3.6823341846466064,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.007027678657323122,
+ "learning_rate": 0.000283051632050831,
+ "loss": 3.7312417030334473,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.0076786368153989315,
+ "learning_rate": 0.00028230786429363046,
+ "loss": 3.775012493133545,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.0067380438558757305,
+ "learning_rate": 0.000281565089694218,
+ "loss": 3.7203941345214844,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.007500625215470791,
+ "learning_rate": 0.00028082330885103516,
+ "loss": 3.7320876121520996,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.006601814646273851,
+ "learning_rate": 0.00028008252236172003,
+ "loss": 3.7289505004882812,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.00755319744348526,
+ "learning_rate": 0.00027934273082311147,
+ "loss": 3.7076311111450195,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.0068661835975945,
+ "learning_rate": 0.0002786039348312459,
+ "loss": 3.6950671672821045,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.007409265730530024,
+ "learning_rate": 0.00027786613498135843,
+ "loss": 3.7798776626586914,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.006714935880154371,
+ "learning_rate": 0.0002771293318678804,
+ "loss": 3.690377712249756,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.007498119492083788,
+ "learning_rate": 0.00027639352608444087,
+ "loss": 3.741222381591797,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.006700210273265839,
+ "learning_rate": 0.0002756587182238656,
+ "loss": 3.7065296173095703,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.008518884889781475,
+ "learning_rate": 0.00027492490887817597,
+ "loss": 3.7010321617126465,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 4.372326374053955,
+ "eval_runtime": 48.97,
+ "eval_samples_per_second": 49.867,
+ "eval_steps_per_second": 1.572,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.007463657297194004,
+ "learning_rate": 0.00027419209863858865,
+ "loss": 3.6893815994262695,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.0069931019097566605,
+ "learning_rate": 0.00027346028809551576,
+ "loss": 3.6620845794677734,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.0076762037351727486,
+ "learning_rate": 0.00027272947783856433,
+ "loss": 3.715946674346924,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.007143040653318167,
+ "learning_rate": 0.0002719996684565341,
+ "loss": 3.67921781539917,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.007577868644148111,
+ "learning_rate": 0.00027127086053741964,
+ "loss": 3.6886801719665527,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.007018061354756355,
+ "learning_rate": 0.0002705430546684085,
+ "loss": 3.673877716064453,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.0072300247848033905,
+ "learning_rate": 0.0002698162514358805,
+ "loss": 3.730193614959717,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.00714140385389328,
+ "learning_rate": 0.00026909045142540843,
+ "loss": 3.6612749099731445,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.0071529485285282135,
+ "learning_rate": 0.0002683656552217553,
+ "loss": 3.641969680786133,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.00726350536569953,
+ "learning_rate": 0.0002676418634088772,
+ "loss": 3.697068452835083,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.007082062773406506,
+ "learning_rate": 0.00026691907656991963,
+ "loss": 3.6732699871063232,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.0069284942001104355,
+ "learning_rate": 0.00026619729528721897,
+ "loss": 3.6701955795288086,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.007283426355570555,
+ "learning_rate": 0.00026547652014230147,
+ "loss": 3.6422624588012695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.006717352196574211,
+ "learning_rate": 0.0002647567517158825,
+ "loss": 3.6600942611694336,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.006842337548732758,
+ "learning_rate": 0.00026403799058786695,
+ "loss": 3.690095901489258,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.006624994333833456,
+ "learning_rate": 0.000263320237337347,
+ "loss": 3.7145729064941406,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.006822563707828522,
+ "learning_rate": 0.0002626034925426036,
+ "loss": 3.734589099884033,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.006343246903270483,
+ "learning_rate": 0.0002618877567811054,
+ "loss": 3.676018238067627,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007226140238344669,
+ "learning_rate": 0.00026117303062950795,
+ "loss": 3.705630302429199,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.0066663846373558044,
+ "learning_rate": 0.0002604593146636522,
+ "loss": 3.707524299621582,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.007017293944954872,
+ "learning_rate": 0.00025974660945856726,
+ "loss": 3.654975652694702,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.006494272034615278,
+ "learning_rate": 0.0002590349155884669,
+ "loss": 3.7084226608276367,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.007430107332766056,
+ "learning_rate": 0.0002583242336267492,
+ "loss": 3.681856393814087,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.0065765841864049435,
+ "learning_rate": 0.00025761456414599807,
+ "loss": 3.7148854732513428,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.00703916372731328,
+ "learning_rate": 0.00025690590771798165,
+ "loss": 3.707977294921875,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.006881422363221645,
+ "learning_rate": 0.0002561982649136517,
+ "loss": 3.6452829837799072,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.006829532328993082,
+ "learning_rate": 0.00025549163630314233,
+ "loss": 3.6812632083892822,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.0069242967292666435,
+ "learning_rate": 0.00025478602245577255,
+ "loss": 3.6699113845825195,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.00673352275043726,
+ "learning_rate": 0.00025408142394004265,
+ "loss": 3.67539644241333,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.006906552705913782,
+ "learning_rate": 0.0002533778413236342,
+ "loss": 3.695683479309082,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007096703629940748,
+ "learning_rate": 0.0002526752751734114,
+ "loss": 3.720625877380371,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006843909155577421,
+ "learning_rate": 0.000251973726055419,
+ "loss": 3.6780426502227783,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.006776071153581142,
+ "learning_rate": 0.00025127319453488277,
+ "loss": 3.702991485595703,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.006684580352157354,
+ "learning_rate": 0.00025057368117620797,
+ "loss": 3.7161166667938232,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006576182786375284,
+ "learning_rate": 0.00024987518654297935,
+ "loss": 3.711947202682495,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.006553746294230223,
+ "learning_rate": 0.0002491777111979629,
+ "loss": 3.699580669403076,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.0063622179441154,
+ "learning_rate": 0.00024848125570310067,
+ "loss": 3.660463809967041,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006369025446474552,
+ "learning_rate": 0.00024778582061951466,
+ "loss": 3.712718963623047,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.0066086845472455025,
+ "learning_rate": 0.00024709140650750426,
+ "loss": 3.7337706089019775,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.006698573008179665,
+ "learning_rate": 0.00024639801392654687,
+ "loss": 3.693399429321289,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006572197191417217,
+ "learning_rate": 0.0002457056434352956,
+ "loss": 3.711644411087036,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.0066736009903252125,
+ "learning_rate": 0.0002450142955915812,
+ "loss": 3.706545352935791,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006779580842703581,
+ "learning_rate": 0.0002443239709524101,
+ "loss": 3.6896963119506836,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.0064375437796115875,
+ "learning_rate": 0.0002436346700739646,
+ "loss": 3.646042823791504,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.006759198382496834,
+ "learning_rate": 0.00024294639351160198,
+ "loss": 3.6894006729125977,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.00644648214802146,
+ "learning_rate": 0.00024225914181985343,
+ "loss": 3.7088422775268555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.00702979601919651,
+ "learning_rate": 0.00024157291555242663,
+ "loss": 3.705204486846924,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006393326912075281,
+ "learning_rate": 0.00024088771526220045,
+ "loss": 3.738778829574585,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.006522264331579208,
+ "learning_rate": 0.00024020354150122884,
+ "loss": 3.7124788761138916,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.006918660830706358,
+ "learning_rate": 0.0002395203948207383,
+ "loss": 3.703273296356201,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.0067405398003757,
+ "learning_rate": 0.00023883827577112786,
+ "loss": 3.735112190246582,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.006900794804096222,
+ "learning_rate": 0.00023815718490196856,
+ "loss": 3.707573413848877,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.006647802423685789,
+ "learning_rate": 0.0002374771227620022,
+ "loss": 3.7194061279296875,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.006786358542740345,
+ "learning_rate": 0.00023679808989914383,
+ "loss": 3.684828758239746,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.006440348923206329,
+ "learning_rate": 0.0002361200868604771,
+ "loss": 3.679628610610962,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.006847103592008352,
+ "learning_rate": 0.0002354431141922572,
+ "loss": 3.7070424556732178,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.00643899803981185,
+ "learning_rate": 0.00023476717243990815,
+ "loss": 3.7474443912506104,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.0070899054408073425,
+ "learning_rate": 0.00023409226214802462,
+ "loss": 3.6964569091796875,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.006639422383159399,
+ "learning_rate": 0.00023341838386036994,
+ "loss": 3.7240355014801025,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.0067732552997767925,
+ "learning_rate": 0.000232745538119875,
+ "loss": 3.691311836242676,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.006991941947489977,
+ "learning_rate": 0.0002320737254686396,
+ "loss": 3.703491449356079,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.006394840776920319,
+ "learning_rate": 0.0002314029464479313,
+ "loss": 3.6963045597076416,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.00672922981902957,
+ "learning_rate": 0.00023073320159818463,
+ "loss": 3.6913256645202637,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.006367729511111975,
+ "learning_rate": 0.00023006449145900022,
+ "loss": 3.712766170501709,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.006370930001139641,
+ "learning_rate": 0.0002293968165691467,
+ "loss": 3.683473587036133,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.007074430584907532,
+ "learning_rate": 0.00022873017746655758,
+ "loss": 3.7364859580993652,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.00689726835116744,
+ "learning_rate": 0.00022806457468833126,
+ "loss": 3.7360360622406006,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.006516663823276758,
+ "learning_rate": 0.0002274000087707323,
+ "loss": 3.7183632850646973,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.006655904930084944,
+ "learning_rate": 0.00022673648024918938,
+ "loss": 3.7091434001922607,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.006546393036842346,
+ "learning_rate": 0.00022607398965829604,
+ "loss": 3.682645320892334,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.006579853594303131,
+ "learning_rate": 0.0002254125375318079,
+ "loss": 3.707092761993408,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.007749224081635475,
+ "learning_rate": 0.00022475212440264524,
+ "loss": 3.6835598945617676,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 4.3733673095703125,
+ "eval_runtime": 54.197,
+ "eval_samples_per_second": 45.058,
+ "eval_steps_per_second": 1.421,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.006405573338270187,
+ "learning_rate": 0.000224092750802892,
+ "loss": 3.7344212532043457,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.006732994224876165,
+ "learning_rate": 0.0002234344172637924,
+ "loss": 3.684095621109009,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.00679958239197731,
+ "learning_rate": 0.00022277712431575434,
+ "loss": 3.697983741760254,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.00670752814039588,
+ "learning_rate": 0.00022212087248834695,
+ "loss": 3.6364173889160156,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.006356659810990095,
+ "learning_rate": 0.00022146566231030071,
+ "loss": 3.642988681793213,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.006624842528253794,
+ "learning_rate": 0.000220811494309506,
+ "loss": 3.6682677268981934,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.006486820988357067,
+ "learning_rate": 0.00022015836901301408,
+ "loss": 3.6703057289123535,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.006477998569607735,
+ "learning_rate": 0.0002195062869470374,
+ "loss": 3.6891942024230957,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.006716342642903328,
+ "learning_rate": 0.0002188552486369454,
+ "loss": 3.6837644577026367,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.006492446642369032,
+ "learning_rate": 0.00021820525460726888,
+ "loss": 3.66949462890625,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.0064978403970599174,
+ "learning_rate": 0.00021755630538169502,
+ "loss": 3.656477451324463,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.006488895509392023,
+ "learning_rate": 0.000216908401483072,
+ "loss": 3.683609962463379,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.0063864924013614655,
+ "learning_rate": 0.00021626154343340316,
+ "loss": 3.6635069847106934,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.006423276849091053,
+ "learning_rate": 0.00021561573175385077,
+ "loss": 3.675956964492798,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.006375681143254042,
+ "learning_rate": 0.00021497096696473364,
+ "loss": 3.662991523742676,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.006157599855214357,
+ "learning_rate": 0.00021432724958552707,
+ "loss": 3.685154438018799,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.0064014773815870285,
+ "learning_rate": 0.0002136845801348629,
+ "loss": 3.6482927799224854,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.006111831869930029,
+ "learning_rate": 0.0002130429591305271,
+ "loss": 3.6458590030670166,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.0062011126428842545,
+ "learning_rate": 0.0002124023870894639,
+ "loss": 3.6651253700256348,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.006603785324841738,
+ "learning_rate": 0.00021176286452776909,
+ "loss": 3.6931943893432617,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.006363562773913145,
+ "learning_rate": 0.00021112439196069516,
+ "loss": 3.7063207626342773,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.006258734501898289,
+ "learning_rate": 0.0002104869699026478,
+ "loss": 3.689711093902588,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.006544198840856552,
+ "learning_rate": 0.00020985059886718659,
+ "loss": 3.6803956031799316,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.006236413959413767,
+ "learning_rate": 0.00020921527936702418,
+ "loss": 3.681662082672119,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.006620924919843674,
+ "learning_rate": 0.00020858101191402557,
+ "loss": 3.6708180904388428,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.006312021054327488,
+ "learning_rate": 0.00020794779701920895,
+ "loss": 3.68668270111084,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.006595146376639605,
+ "learning_rate": 0.00020731563519274407,
+ "loss": 3.709224224090576,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.006226524710655212,
+ "learning_rate": 0.00020668452694395234,
+ "loss": 3.688267707824707,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.006482335273176432,
+ "learning_rate": 0.00020605447278130616,
+ "loss": 3.6906940937042236,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.006389188580214977,
+ "learning_rate": 0.00020542547321242892,
+ "loss": 3.6813313961029053,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.006554862484335899,
+ "learning_rate": 0.00020479752874409457,
+ "loss": 3.645631790161133,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.00646898802369833,
+ "learning_rate": 0.0002041706398822257,
+ "loss": 3.691580295562744,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.006872115656733513,
+ "learning_rate": 0.0002035448071318961,
+ "loss": 3.716787815093994,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.006251885090023279,
+ "learning_rate": 0.00020292003099732744,
+ "loss": 3.666078567504883,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.006771737243980169,
+ "learning_rate": 0.00020229631198189084,
+ "loss": 3.621438503265381,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.00606488948687911,
+ "learning_rate": 0.0002016736505881058,
+ "loss": 3.6710901260375977,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.006203590892255306,
+ "learning_rate": 0.00020105204731763832,
+ "loss": 3.6667685508728027,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.006625751499086618,
+ "learning_rate": 0.00020043150267130412,
+ "loss": 3.648806095123291,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.006358640734106302,
+ "learning_rate": 0.000199812017149064,
+ "loss": 3.7056827545166016,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.006539507769048214,
+ "learning_rate": 0.00019919359125002647,
+ "loss": 3.6906323432922363,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.007050348911434412,
+ "learning_rate": 0.00019857622547244627,
+ "loss": 3.6898245811462402,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.006209646817296743,
+ "learning_rate": 0.00019795992031372358,
+ "loss": 3.6794214248657227,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.0064559150487184525,
+ "learning_rate": 0.00019734467627040487,
+ "loss": 3.6937336921691895,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.006774613168090582,
+ "learning_rate": 0.00019673049383818025,
+ "loss": 3.7005720138549805,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.0061691561713814735,
+ "learning_rate": 0.00019611737351188642,
+ "loss": 3.7170605659484863,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.0070815603248775005,
+ "learning_rate": 0.0001955053157855027,
+ "loss": 3.722456693649292,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.006411026697605848,
+ "learning_rate": 0.00019489432115215308,
+ "loss": 3.669548511505127,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.006817718036472797,
+ "learning_rate": 0.00019428439010410492,
+ "loss": 3.705488681793213,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.006276566535234451,
+ "learning_rate": 0.0001936755231327688,
+ "loss": 3.7110562324523926,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.006713696289807558,
+ "learning_rate": 0.0001930677207286977,
+ "loss": 3.7030649185180664,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.006331400014460087,
+ "learning_rate": 0.00019246098338158676,
+ "loss": 3.6906652450561523,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.007149284239858389,
+ "learning_rate": 0.00019185531158027384,
+ "loss": 3.7055468559265137,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.006395523902028799,
+ "learning_rate": 0.00019125070581273734,
+ "loss": 3.72355318069458,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.007234807591885328,
+ "learning_rate": 0.0001906471665660978,
+ "loss": 3.6685009002685547,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.006079623010009527,
+ "learning_rate": 0.00019004469432661478,
+ "loss": 3.7051684856414795,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.00699756620451808,
+ "learning_rate": 0.0001894432895796907,
+ "loss": 3.6713948249816895,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.006799036171287298,
+ "learning_rate": 0.00018884295280986531,
+ "loss": 3.7098617553710938,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.006591275800019503,
+ "learning_rate": 0.0001882436845008197,
+ "loss": 3.685445785522461,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.006709378212690353,
+ "learning_rate": 0.0001876454851353733,
+ "loss": 3.6773548126220703,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.00651182746514678,
+ "learning_rate": 0.0001870483551954847,
+ "loss": 3.6529507637023926,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.006494257599115372,
+ "learning_rate": 0.00018645229516225093,
+ "loss": 3.692331552505493,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.006487370003014803,
+ "learning_rate": 0.00018585730551590619,
+ "loss": 3.679516315460205,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.006413135677576065,
+ "learning_rate": 0.00018526338673582307,
+ "loss": 3.707998752593994,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.006509249564260244,
+ "learning_rate": 0.00018467053930051156,
+ "loss": 3.7103686332702637,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.006331913638859987,
+ "learning_rate": 0.0001840787636876177,
+ "loss": 3.6807055473327637,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.006528888829052448,
+ "learning_rate": 0.00018348806037392473,
+ "loss": 3.7196922302246094,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.006212258245795965,
+ "learning_rate": 0.0001828984298353515,
+ "loss": 3.716118335723877,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.006641864310950041,
+ "learning_rate": 0.000182309872546953,
+ "loss": 3.6726081371307373,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.006433268543332815,
+ "learning_rate": 0.0001817223889829186,
+ "loss": 3.690727710723877,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.00644147302955389,
+ "learning_rate": 0.00018113597961657364,
+ "loss": 3.6998791694641113,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.006477390881627798,
+ "learning_rate": 0.00018055064492037722,
+ "loss": 3.700925350189209,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.0077974810265004635,
+ "learning_rate": 0.00017996638536592322,
+ "loss": 3.735599994659424,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 4.376750946044922,
+ "eval_runtime": 52.3576,
+ "eval_samples_per_second": 46.641,
+ "eval_steps_per_second": 1.471,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.006609050091356039,
+ "learning_rate": 0.00017938320142393897,
+ "loss": 3.6405727863311768,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.006470015738159418,
+ "learning_rate": 0.00017880109356428522,
+ "loss": 3.6639411449432373,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.006597487255930901,
+ "learning_rate": 0.00017822006225595596,
+ "loss": 3.6797537803649902,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.0064637972973287106,
+ "learning_rate": 0.0001776401079670772,
+ "loss": 3.6798789501190186,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.006872601341456175,
+ "learning_rate": 0.00017706123116490775,
+ "loss": 3.649874210357666,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.0064028468914330006,
+ "learning_rate": 0.00017648343231583825,
+ "loss": 3.671036720275879,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.006735685281455517,
+ "learning_rate": 0.00017590671188539084,
+ "loss": 3.6818368434906006,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.006652928423136473,
+ "learning_rate": 0.00017533107033821895,
+ "loss": 3.660362720489502,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.006564146373420954,
+ "learning_rate": 0.00017475650813810598,
+ "loss": 3.6444082260131836,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.006476754322648048,
+ "learning_rate": 0.00017418302574796716,
+ "loss": 3.6836254596710205,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.007108788937330246,
+ "learning_rate": 0.0001736106236298462,
+ "loss": 3.6424341201782227,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.006570066325366497,
+ "learning_rate": 0.00017303930224491735,
+ "loss": 3.663379192352295,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.0066614942625164986,
+ "learning_rate": 0.00017246906205348395,
+ "loss": 3.6918721199035645,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.00704583153128624,
+ "learning_rate": 0.00017189990351497824,
+ "loss": 3.692559242248535,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.006617981940507889,
+ "learning_rate": 0.00017133182708796126,
+ "loss": 3.6591439247131348,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.006916946265846491,
+ "learning_rate": 0.00017076483323012127,
+ "loss": 3.7165699005126953,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.006832561455667019,
+ "learning_rate": 0.00017019892239827534,
+ "loss": 3.7023603916168213,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.00654558464884758,
+ "learning_rate": 0.00016963409504836777,
+ "loss": 3.6702237129211426,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.007000794634222984,
+ "learning_rate": 0.0001690703516354698,
+ "loss": 3.69088077545166,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.006473604589700699,
+ "learning_rate": 0.00016850769261377856,
+ "loss": 3.6641716957092285,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.006787394639104605,
+ "learning_rate": 0.000167946118436619,
+ "loss": 3.6577839851379395,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.006220053881406784,
+ "learning_rate": 0.00016738562955644144,
+ "loss": 3.6735315322875977,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.006513087544590235,
+ "learning_rate": 0.00016682622642482113,
+ "loss": 3.624807119369507,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.006552594713866711,
+ "learning_rate": 0.0001662679094924592,
+ "loss": 3.679218053817749,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.006950687617063522,
+ "learning_rate": 0.0001657106792091814,
+ "loss": 3.6864023208618164,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.006092227529734373,
+ "learning_rate": 0.0001651545360239387,
+ "loss": 3.693777084350586,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.006924188230186701,
+ "learning_rate": 0.0001645994803848046,
+ "loss": 3.68534779548645,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.0067315176129341125,
+ "learning_rate": 0.00016404551273897826,
+ "loss": 3.632850170135498,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.007006707601249218,
+ "learning_rate": 0.00016349263353278146,
+ "loss": 3.6664111614227295,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.0067031108774244785,
+ "learning_rate": 0.00016294084321165846,
+ "loss": 3.6863315105438232,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.006502983160316944,
+ "learning_rate": 0.00016239014222017682,
+ "loss": 3.678041458129883,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.006409413646906614,
+ "learning_rate": 0.00016184053100202665,
+ "loss": 3.692004919052124,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.006251887418329716,
+ "learning_rate": 0.0001612920100000198,
+ "loss": 3.64532470703125,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.0065657407976686954,
+ "learning_rate": 0.00016074457965608942,
+ "loss": 3.638624429702759,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.006246988195925951,
+ "learning_rate": 0.00016019824041129004,
+ "loss": 3.668254852294922,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.0065050385892391205,
+ "learning_rate": 0.0001596529927057984,
+ "loss": 3.6901469230651855,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.006240491755306721,
+ "learning_rate": 0.00015910883697890977,
+ "loss": 3.6796071529388428,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.006674987263977528,
+ "learning_rate": 0.00015856577366904075,
+ "loss": 3.7001953125,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.006260441616177559,
+ "learning_rate": 0.00015802380321372776,
+ "loss": 3.6646556854248047,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.006696469150483608,
+ "learning_rate": 0.00015748292604962687,
+ "loss": 3.6791043281555176,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.006716041825711727,
+ "learning_rate": 0.00015694314261251245,
+ "loss": 3.710442066192627,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.006856406573206186,
+ "learning_rate": 0.0001564044533372785,
+ "loss": 3.6615681648254395,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.0064330073073506355,
+ "learning_rate": 0.00015586685865793748,
+ "loss": 3.707085609436035,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.006748909596353769,
+ "learning_rate": 0.00015533035900761955,
+ "loss": 3.6814656257629395,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.006306119728833437,
+ "learning_rate": 0.00015479495481857282,
+ "loss": 3.6595420837402344,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.0061928280629217625,
+ "learning_rate": 0.00015426064652216255,
+ "loss": 3.686880111694336,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.006499494891613722,
+ "learning_rate": 0.00015372743454887173,
+ "loss": 3.6901583671569824,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.005914116278290749,
+ "learning_rate": 0.00015319531932829917,
+ "loss": 3.699162006378174,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.006499269045889378,
+ "learning_rate": 0.00015266430128916075,
+ "loss": 3.6740684509277344,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.006433961447328329,
+ "learning_rate": 0.000152134380859288,
+ "loss": 3.684006929397583,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.006057603284716606,
+ "learning_rate": 0.00015160555846562826,
+ "loss": 3.6612191200256348,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.006337927188724279,
+ "learning_rate": 0.00015107783453424456,
+ "loss": 3.6981983184814453,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.006348243448883295,
+ "learning_rate": 0.00015055120949031365,
+ "loss": 3.6667895317077637,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.0060755726881325245,
+ "learning_rate": 0.00015002568375812876,
+ "loss": 3.6844797134399414,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.006059548351913691,
+ "learning_rate": 0.0001495012577610959,
+ "loss": 3.6676461696624756,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.006432189140468836,
+ "learning_rate": 0.00014897793192173602,
+ "loss": 3.7036523818969727,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.006043825298547745,
+ "learning_rate": 0.00014845570666168249,
+ "loss": 3.7043371200561523,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.0066383229568600655,
+ "learning_rate": 0.0001479345824016837,
+ "loss": 3.6830334663391113,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.006384526379406452,
+ "learning_rate": 0.00014741455956160005,
+ "loss": 3.6746675968170166,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.006477308925241232,
+ "learning_rate": 0.0001468956385604038,
+ "loss": 3.6788225173950195,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.006130026653409004,
+ "learning_rate": 0.00014637781981618054,
+ "loss": 3.6821796894073486,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.006257957778871059,
+ "learning_rate": 0.00014586110374612777,
+ "loss": 3.683966636657715,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.006265794392675161,
+ "learning_rate": 0.00014534549076655425,
+ "loss": 3.691389560699463,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.006294097751379013,
+ "learning_rate": 0.0001448309812928794,
+ "loss": 3.659594774246216,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.006432943977415562,
+ "learning_rate": 0.00014431757573963504,
+ "loss": 3.665633201599121,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.006102072540670633,
+ "learning_rate": 0.0001438052745204626,
+ "loss": 3.6348228454589844,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.00636800192296505,
+ "learning_rate": 0.00014329407804811333,
+ "loss": 3.624040365219116,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.006060190033167601,
+ "learning_rate": 0.00014278398673444966,
+ "loss": 3.6986045837402344,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.0061101666651666164,
+ "learning_rate": 0.00014227500099044254,
+ "loss": 3.6816675662994385,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.006048897281289101,
+ "learning_rate": 0.00014176712122617293,
+ "loss": 3.6739861965179443,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.005954945459961891,
+ "learning_rate": 0.00014126034785082997,
+ "loss": 3.679354190826416,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.007545145694166422,
+ "learning_rate": 0.0001407546812727119,
+ "loss": 3.7448782920837402,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 4.377684116363525,
+ "eval_runtime": 51.0591,
+ "eval_samples_per_second": 47.827,
+ "eval_steps_per_second": 1.508,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.00643246341496706,
+ "learning_rate": 0.00014025012189922562,
+ "loss": 3.6544783115386963,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.006309461314231157,
+ "learning_rate": 0.00013974667013688495,
+ "loss": 3.701963186264038,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.006177262403070927,
+ "learning_rate": 0.00013924432639131215,
+ "loss": 3.649641275405884,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.006319771986454725,
+ "learning_rate": 0.00013874309106723636,
+ "loss": 3.67417573928833,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.006548433564603329,
+ "learning_rate": 0.00013824296456849426,
+ "loss": 3.6257686614990234,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.006148583721369505,
+ "learning_rate": 0.00013774394729802804,
+ "loss": 3.675339460372925,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.0063438634388148785,
+ "learning_rate": 0.00013724603965788716,
+ "loss": 3.666316270828247,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.006417148746550083,
+ "learning_rate": 0.00013674924204922732,
+ "loss": 3.6692237854003906,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.0062498473562300205,
+ "learning_rate": 0.00013625355487230872,
+ "loss": 3.70363187789917,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.006249656435102224,
+ "learning_rate": 0.00013575897852649796,
+ "loss": 3.691108226776123,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.006240040063858032,
+ "learning_rate": 0.0001352655134102657,
+ "loss": 3.6781342029571533,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.006186297629028559,
+ "learning_rate": 0.00013477315992118885,
+ "loss": 3.644334316253662,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.006283515132963657,
+ "learning_rate": 0.00013428191845594675,
+ "loss": 3.6691460609436035,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.0064705004915595055,
+ "learning_rate": 0.00013379178941032427,
+ "loss": 3.64168119430542,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.006210929248481989,
+ "learning_rate": 0.00013330277317920959,
+ "loss": 3.66428804397583,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.006095333956182003,
+ "learning_rate": 0.00013281487015659395,
+ "loss": 3.6205215454101562,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.006364993751049042,
+ "learning_rate": 0.00013232808073557244,
+ "loss": 3.662804126739502,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.006166066508740187,
+ "learning_rate": 0.00013184240530834173,
+ "loss": 3.702338218688965,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.006147229578346014,
+ "learning_rate": 0.00013135784426620282,
+ "loss": 3.660149097442627,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.006086066365242004,
+ "learning_rate": 0.00013087439799955674,
+ "loss": 3.652472972869873,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.006371276918798685,
+ "learning_rate": 0.00013039206689790815,
+ "loss": 3.6471242904663086,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.005952878389507532,
+ "learning_rate": 0.000129910851349862,
+ "loss": 3.6450729370117188,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.006385558284819126,
+ "learning_rate": 0.00012943075174312536,
+ "loss": 3.6730473041534424,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.005896324757486582,
+ "learning_rate": 0.00012895176846450582,
+ "loss": 3.6646409034729004,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.006053723394870758,
+ "learning_rate": 0.00012847390189991121,
+ "loss": 3.6568682193756104,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.005929332692176104,
+ "learning_rate": 0.00012799715243435027,
+ "loss": 3.7030997276306152,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.006153058726340532,
+ "learning_rate": 0.00012752152045193116,
+ "loss": 3.6949849128723145,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.006193236447870731,
+ "learning_rate": 0.0001270470063358624,
+ "loss": 3.6040501594543457,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.005849961191415787,
+ "learning_rate": 0.0001265736104684509,
+ "loss": 3.602717399597168,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.0060405549593269825,
+ "learning_rate": 0.00012610133323110356,
+ "loss": 3.668717861175537,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.0062002381309866905,
+ "learning_rate": 0.00012563017500432565,
+ "loss": 3.6333389282226562,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.006093314848840237,
+ "learning_rate": 0.00012516013616772065,
+ "loss": 3.662696361541748,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.005765238776803017,
+ "learning_rate": 0.00012469121709999026,
+ "loss": 3.6843791007995605,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.005989165045320988,
+ "learning_rate": 0.00012422341817893448,
+ "loss": 3.6904735565185547,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.006150389555841684,
+ "learning_rate": 0.00012375673978145022,
+ "loss": 3.6614508628845215,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.005898299627006054,
+ "learning_rate": 0.00012329118228353178,
+ "loss": 3.7193222045898438,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.0058402069844305515,
+ "learning_rate": 0.00012282674606027016,
+ "loss": 3.669564723968506,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.006015103310346603,
+ "learning_rate": 0.00012236343148585413,
+ "loss": 3.686734914779663,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.005990280769765377,
+ "learning_rate": 0.00012190123893356703,
+ "loss": 3.664945125579834,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.006110248621553183,
+ "learning_rate": 0.00012144016877578942,
+ "loss": 3.6546223163604736,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.006006988696753979,
+ "learning_rate": 0.000120980221383997,
+ "loss": 3.669175386428833,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.006008700001984835,
+ "learning_rate": 0.00012052139712876141,
+ "loss": 3.672668695449829,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.005862198304384947,
+ "learning_rate": 0.00012006369637974875,
+ "loss": 3.6824991703033447,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.005960221868008375,
+ "learning_rate": 0.00011960711950572005,
+ "loss": 3.6258697509765625,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.005813480354845524,
+ "learning_rate": 0.00011915166687453174,
+ "loss": 3.6463029384613037,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.006002446636557579,
+ "learning_rate": 0.0001186973388531332,
+ "loss": 3.721970558166504,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.006029803771525621,
+ "learning_rate": 0.00011824413580756844,
+ "loss": 3.685335159301758,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.005933200940489769,
+ "learning_rate": 0.0001177920581029752,
+ "loss": 3.6834492683410645,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.005942870397120714,
+ "learning_rate": 0.00011734110610358423,
+ "loss": 3.670149326324463,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.005972386337816715,
+ "learning_rate": 0.00011689128017271901,
+ "loss": 3.6798110008239746,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.005964178126305342,
+ "learning_rate": 0.00011644258067279662,
+ "loss": 3.6620044708251953,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.0059594810009002686,
+ "learning_rate": 0.00011599500796532605,
+ "loss": 3.683058738708496,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.005901646800339222,
+ "learning_rate": 0.00011554856241090844,
+ "loss": 3.6791443824768066,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.006178564857691526,
+ "learning_rate": 0.00011510324436923708,
+ "loss": 3.6620421409606934,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.005973632447421551,
+ "learning_rate": 0.0001146590541990962,
+ "loss": 3.6821961402893066,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.005914160516113043,
+ "learning_rate": 0.00011421599225836265,
+ "loss": 3.6588668823242188,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.005887249484658241,
+ "learning_rate": 0.00011377405890400266,
+ "loss": 3.683187484741211,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.006068149581551552,
+ "learning_rate": 0.00011333325449207415,
+ "loss": 3.6633477210998535,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.005748802796006203,
+ "learning_rate": 0.00011289357937772547,
+ "loss": 3.676607608795166,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.006032158620655537,
+ "learning_rate": 0.00011245503391519489,
+ "loss": 3.652369499206543,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.00615714630112052,
+ "learning_rate": 0.00011201761845781049,
+ "loss": 3.7210354804992676,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.0059432960115373135,
+ "learning_rate": 0.00011158133335799003,
+ "loss": 3.640336036682129,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.0058724163100123405,
+ "learning_rate": 0.00011114617896724026,
+ "loss": 3.638444185256958,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.006073266267776489,
+ "learning_rate": 0.00011071215563615752,
+ "loss": 3.680121421813965,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.005904203746467829,
+ "learning_rate": 0.00011027926371442662,
+ "loss": 3.6883950233459473,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.005889971274882555,
+ "learning_rate": 0.00010984750355082045,
+ "loss": 3.6875810623168945,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.005878197029232979,
+ "learning_rate": 0.00010941687549320068,
+ "loss": 3.6341497898101807,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.005862570833414793,
+ "learning_rate": 0.00010898737988851677,
+ "loss": 3.648092746734619,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.0058378661051392555,
+ "learning_rate": 0.00010855901708280482,
+ "loss": 3.627786636352539,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0057954080402851105,
+ "learning_rate": 0.0001081317874211895,
+ "loss": 3.636265754699707,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.005763230379670858,
+ "learning_rate": 0.00010770569124788201,
+ "loss": 3.6991324424743652,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.007435750216245651,
+ "learning_rate": 0.00010728072890618029,
+ "loss": 3.698052406311035,
+ "step": 3240
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.3813260125995008e+18,
+ "train_batch_size": 16,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/baseline/checkpoint-3240/training_args.bin b/runs/baseline/checkpoint-3240/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7
--- /dev/null
+++ b/runs/baseline/checkpoint-3240/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97
+size 5265
diff --git a/runs/baseline/checkpoint-360/chat_template.jinja b/runs/baseline/checkpoint-360/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/baseline/checkpoint-360/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/baseline/checkpoint-360/config.json b/runs/baseline/checkpoint-360/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/baseline/checkpoint-360/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/baseline/checkpoint-360/generation_config.json b/runs/baseline/checkpoint-360/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/baseline/checkpoint-360/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/baseline/checkpoint-360/model.safetensors b/runs/baseline/checkpoint-360/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..5db2a911183563f3ac12920bdb93a1e3b04d2886
--- /dev/null
+++ b/runs/baseline/checkpoint-360/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e278725cdeece35f0bd1ccc41b8ceef8ec080ca3c6adb12629b41bab70293699
+size 583356232
diff --git a/runs/baseline/checkpoint-360/optimizer.pt b/runs/baseline/checkpoint-360/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..5225a05ef9671f4864b140f4c2629b84e4791604
--- /dev/null
+++ b/runs/baseline/checkpoint-360/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4f656fc0021dd1568063069ead3d1c685d1d893d5bd6dc43f598982bd9201f93
+size 1166825803
diff --git a/runs/baseline/checkpoint-360/rng_state_0.pth b/runs/baseline/checkpoint-360/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3a2bb760d0b0b50f4b7b6980626fda172a127f68
--- /dev/null
+++ b/runs/baseline/checkpoint-360/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9627a7b92eeed978894313d8642baf0ecdcd808f2102543d8b58136102ad5997
+size 14917
diff --git a/runs/baseline/checkpoint-360/rng_state_1.pth b/runs/baseline/checkpoint-360/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..7b307a1ab4859d8d838e7d20293e5eb15985549a
--- /dev/null
+++ b/runs/baseline/checkpoint-360/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0d79c621c2f337ae714092486c08d5a224dca6ddc4285fb0014e9b648d0c9cae
+size 14917
diff --git a/runs/baseline/checkpoint-360/scheduler.pt b/runs/baseline/checkpoint-360/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..3acc96860dc2d45c77cec0d2f55bac5383c2a389
--- /dev/null
+++ b/runs/baseline/checkpoint-360/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9c2cb27444ef7fe125f0634d6f0887d628ac22fdc5ea5d28ad2b2519c38009cf
+size 1465
diff --git a/runs/baseline/checkpoint-360/tokenizer.json b/runs/baseline/checkpoint-360/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/baseline/checkpoint-360/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/baseline/checkpoint-360/tokenizer_config.json b/runs/baseline/checkpoint-360/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/baseline/checkpoint-360/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/baseline/checkpoint-360/trainer_state.json b/runs/baseline/checkpoint-360/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..a8bbe99d48af8155557780ea09d06ee9992b0a43
--- /dev/null
+++ b/runs/baseline/checkpoint-360/trainer_state.json
@@ -0,0 +1,2586 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 5.0,
+ "eval_steps": 500,
+ "global_step": 360,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.12165174633264542,
+ "learning_rate": 0.0,
+ "loss": 12.027063369750977,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.12109383195638657,
+ "learning_rate": 3e-05,
+ "loss": 12.028635025024414,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.12690064311027527,
+ "learning_rate": 6e-05,
+ "loss": 11.894524574279785,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.1537286937236786,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.672262191772461,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.11778244376182556,
+ "learning_rate": 0.00012,
+ "loss": 11.433671951293945,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.10874935984611511,
+ "learning_rate": 0.00015000000000000001,
+ "loss": 11.270309448242188,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.10641921311616898,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 11.155481338500977,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.1048506498336792,
+ "learning_rate": 0.00021000000000000004,
+ "loss": 11.066675186157227,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.10546834021806717,
+ "learning_rate": 0.00024,
+ "loss": 10.952295303344727,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.10614268481731415,
+ "learning_rate": 0.00027,
+ "loss": 10.811386108398438,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10529010742902756,
+ "learning_rate": 0.00030000000000000003,
+ "loss": 10.657712936401367,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.1051444336771965,
+ "learning_rate": 0.00033,
+ "loss": 10.47591781616211,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10420846194028854,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 10.280141830444336,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10323625057935715,
+ "learning_rate": 0.00039000000000000005,
+ "loss": 10.076366424560547,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10255401581525803,
+ "learning_rate": 0.00042000000000000007,
+ "loss": 9.856939315795898,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10151027888059616,
+ "learning_rate": 0.00045,
+ "loss": 9.632134437561035,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.1006326898932457,
+ "learning_rate": 0.00048,
+ "loss": 9.401637077331543,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.09820562601089478,
+ "learning_rate": 0.00051,
+ "loss": 9.177400588989258,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.0956626906991005,
+ "learning_rate": 0.00054,
+ "loss": 8.95518684387207,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.0922091156244278,
+ "learning_rate": 0.00057,
+ "loss": 8.75066089630127,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.08593932539224625,
+ "learning_rate": 0.0006000000000000001,
+ "loss": 8.557976722717285,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.08014766871929169,
+ "learning_rate": 0.00063,
+ "loss": 8.360604286193848,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.07381244748830795,
+ "learning_rate": 0.00066,
+ "loss": 8.210375785827637,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.06455052644014359,
+ "learning_rate": 0.0006900000000000001,
+ "loss": 8.07168960571289,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.054115474224090576,
+ "learning_rate": 0.0007199999999999999,
+ "loss": 7.960069179534912,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.043575264513492584,
+ "learning_rate": 0.00075,
+ "loss": 7.850685119628906,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.02976679615676403,
+ "learning_rate": 0.0007800000000000001,
+ "loss": 7.824377059936523,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.020552165806293488,
+ "learning_rate": 0.0008100000000000001,
+ "loss": 7.7695417404174805,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.018530743196606636,
+ "learning_rate": 0.0008400000000000001,
+ "loss": 7.782247543334961,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.02716820128262043,
+ "learning_rate": 0.00087,
+ "loss": 7.754945278167725,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.03638409823179245,
+ "learning_rate": 0.0009,
+ "loss": 7.799381256103516,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.041579652577638626,
+ "learning_rate": 0.00093,
+ "loss": 7.776629447937012,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.0304960235953331,
+ "learning_rate": 0.00096,
+ "loss": 7.749439239501953,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.053728822618722916,
+ "learning_rate": 0.00099,
+ "loss": 7.745059490203857,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.036322493106126785,
+ "learning_rate": 0.00102,
+ "loss": 7.752074718475342,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.03232959285378456,
+ "learning_rate": 0.00105,
+ "loss": 7.704257011413574,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.03246184438467026,
+ "learning_rate": 0.00108,
+ "loss": 7.667045593261719,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.036655545234680176,
+ "learning_rate": 0.00111,
+ "loss": 7.623445987701416,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.021042725071310997,
+ "learning_rate": 0.00114,
+ "loss": 7.582821846008301,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.022982975468039513,
+ "learning_rate": 0.00117,
+ "loss": 7.58036994934082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.024781864136457443,
+ "learning_rate": 0.0012000000000000001,
+ "loss": 7.5903000831604,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.03667089343070984,
+ "learning_rate": 0.00123,
+ "loss": 7.588233947753906,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.04083017632365227,
+ "learning_rate": 0.00126,
+ "loss": 7.548343658447266,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.02585385926067829,
+ "learning_rate": 0.00129,
+ "loss": 7.604833602905273,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.018793931230902672,
+ "learning_rate": 0.00132,
+ "loss": 7.56705379486084,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.025538936257362366,
+ "learning_rate": 0.00135,
+ "loss": 7.562107086181641,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.02819318138062954,
+ "learning_rate": 0.0013800000000000002,
+ "loss": 7.517213821411133,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.016239255666732788,
+ "learning_rate": 0.00141,
+ "loss": 7.4780755043029785,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.024470701813697815,
+ "learning_rate": 0.0014399999999999999,
+ "loss": 7.505887985229492,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.015499847941100597,
+ "learning_rate": 0.00147,
+ "loss": 7.484469890594482,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.017785200849175453,
+ "learning_rate": 0.0015,
+ "loss": 7.481822967529297,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.015952805057168007,
+ "learning_rate": 0.0015300000000000001,
+ "loss": 7.484804153442383,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.015771260485053062,
+ "learning_rate": 0.0015600000000000002,
+ "loss": 7.4495439529418945,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.024370530620217323,
+ "learning_rate": 0.00159,
+ "loss": 7.471208572387695,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.018173737451434135,
+ "learning_rate": 0.0016200000000000001,
+ "loss": 7.438447952270508,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.020564094185829163,
+ "learning_rate": 0.0016500000000000002,
+ "loss": 7.437922954559326,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.016154944896697998,
+ "learning_rate": 0.0016800000000000003,
+ "loss": 7.429378032684326,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.016817007213830948,
+ "learning_rate": 0.00171,
+ "loss": 7.406428337097168,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.03460947796702385,
+ "learning_rate": 0.00174,
+ "loss": 7.393331527709961,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.15410488843917847,
+ "learning_rate": 0.0017699999999999999,
+ "loss": 7.645856857299805,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.07443148642778397,
+ "learning_rate": 0.0018,
+ "loss": 7.549508094787598,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.0507572740316391,
+ "learning_rate": 0.00183,
+ "loss": 7.517921447753906,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.09682746231555939,
+ "learning_rate": 0.00186,
+ "loss": 7.543978691101074,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.2571517527103424,
+ "learning_rate": 0.00189,
+ "loss": 7.768949508666992,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.0317784883081913,
+ "learning_rate": 0.00192,
+ "loss": 7.509063720703125,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.0353970006108284,
+ "learning_rate": 0.0019500000000000001,
+ "loss": 7.522322654724121,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.03663994371891022,
+ "learning_rate": 0.00198,
+ "loss": 7.504723072052002,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.025481393560767174,
+ "learning_rate": 0.00201,
+ "loss": 7.452597141265869,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.03099915198981762,
+ "learning_rate": 0.00204,
+ "loss": 7.455441474914551,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.022921698167920113,
+ "learning_rate": 0.00207,
+ "loss": 7.4454827308654785,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.032388731837272644,
+ "learning_rate": 0.0021,
+ "loss": 7.467103958129883,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.022428609430789948,
+ "learning_rate": 0.00213,
+ "loss": 7.467540264129639,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.469719409942627,
+ "eval_runtime": 50.9714,
+ "eval_samples_per_second": 47.909,
+ "eval_steps_per_second": 1.511,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.017480911687016487,
+ "learning_rate": 0.00216,
+ "loss": 7.455810546875,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.016675131395459175,
+ "learning_rate": 0.00219,
+ "loss": 7.474848747253418,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.017450228333473206,
+ "learning_rate": 0.00222,
+ "loss": 7.442611217498779,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.016823675483465195,
+ "learning_rate": 0.0022500000000000003,
+ "loss": 7.454628944396973,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.016717543825507164,
+ "learning_rate": 0.00228,
+ "loss": 7.425317764282227,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.01644429750740528,
+ "learning_rate": 0.00231,
+ "loss": 7.4102864265441895,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.013104400597512722,
+ "learning_rate": 0.00234,
+ "loss": 7.4559173583984375,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.01175684668123722,
+ "learning_rate": 0.00237,
+ "loss": 7.438898086547852,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.013736684806644917,
+ "learning_rate": 0.0024000000000000002,
+ "loss": 7.449005603790283,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01176756713539362,
+ "learning_rate": 0.0024300000000000003,
+ "loss": 7.428421974182129,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.011860918253660202,
+ "learning_rate": 0.00246,
+ "loss": 7.411491394042969,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.010093477554619312,
+ "learning_rate": 0.00249,
+ "loss": 7.39311408996582,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.009453288279473782,
+ "learning_rate": 0.00252,
+ "loss": 7.387656211853027,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.010311692021787167,
+ "learning_rate": 0.00255,
+ "loss": 7.401076316833496,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.010053195990622044,
+ "learning_rate": 0.00258,
+ "loss": 7.39655876159668,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.009758037514984608,
+ "learning_rate": 0.00261,
+ "loss": 7.423122882843018,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008394946344196796,
+ "learning_rate": 0.00264,
+ "loss": 7.383601188659668,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.008457187563180923,
+ "learning_rate": 0.00267,
+ "loss": 7.400934219360352,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.010601334273815155,
+ "learning_rate": 0.0027,
+ "loss": 7.400241851806641,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.009009373374283314,
+ "learning_rate": 0.0027300000000000002,
+ "loss": 7.377348899841309,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.00737508200109005,
+ "learning_rate": 0.0027600000000000003,
+ "loss": 7.385881423950195,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.0075711836107075214,
+ "learning_rate": 0.0027900000000000004,
+ "loss": 7.348570823669434,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.00846901722252369,
+ "learning_rate": 0.00282,
+ "loss": 7.364287853240967,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.007160635199397802,
+ "learning_rate": 0.00285,
+ "loss": 7.3881635665893555,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.00682067358866334,
+ "learning_rate": 0.0028799999999999997,
+ "loss": 7.348663330078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.009108101949095726,
+ "learning_rate": 0.00291,
+ "loss": 7.356302261352539,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.015828927978873253,
+ "learning_rate": 0.00294,
+ "loss": 7.347576141357422,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.009683096781373024,
+ "learning_rate": 0.00297,
+ "loss": 7.323967933654785,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.010833270847797394,
+ "learning_rate": 0.003,
+ "loss": 7.3139495849609375,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.014905400574207306,
+ "learning_rate": 0.002999999401781161,
+ "loss": 7.302030086517334,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.023674853146076202,
+ "learning_rate": 0.002999997607125127,
+ "loss": 7.311811923980713,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.050460800528526306,
+ "learning_rate": 0.002999994616033342,
+ "loss": 7.41972541809082,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.02193174883723259,
+ "learning_rate": 0.002999990428508218,
+ "loss": 7.294468879699707,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.036261629313230515,
+ "learning_rate": 0.0029999850445531283,
+ "loss": 7.295568466186523,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.02734929881989956,
+ "learning_rate": 0.002999978464172409,
+ "loss": 7.309239864349365,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.017884107306599617,
+ "learning_rate": 0.002999970687371363,
+ "loss": 7.2427592277526855,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01700308918952942,
+ "learning_rate": 0.0029999617141562573,
+ "loss": 7.241730213165283,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.0209935512393713,
+ "learning_rate": 0.0029999515445343194,
+ "loss": 7.262085437774658,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.023207105696201324,
+ "learning_rate": 0.0029999401785137437,
+ "loss": 7.25450325012207,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.01907353848218918,
+ "learning_rate": 0.002999927616103687,
+ "loss": 7.23149299621582,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.016149666160345078,
+ "learning_rate": 0.0029999138573142715,
+ "loss": 7.210232734680176,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.017513928934931755,
+ "learning_rate": 0.002999898902156582,
+ "loss": 7.218451976776123,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.015002547763288021,
+ "learning_rate": 0.002999882750642667,
+ "loss": 7.169316291809082,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.014233081601560116,
+ "learning_rate": 0.0029998654027855406,
+ "loss": 7.194825172424316,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.01283260714262724,
+ "learning_rate": 0.0029998468585991783,
+ "loss": 7.15669059753418,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.014416799880564213,
+ "learning_rate": 0.0029998271180985223,
+ "loss": 7.143369674682617,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.014337058179080486,
+ "learning_rate": 0.0029998061812994757,
+ "loss": 7.127734184265137,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.016406210139393806,
+ "learning_rate": 0.002999784048218908,
+ "loss": 7.086504936218262,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.01680421642959118,
+ "learning_rate": 0.002999760718874652,
+ "loss": 7.089420318603516,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.010665548034012318,
+ "learning_rate": 0.0029997361932855016,
+ "loss": 7.102463722229004,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013479137793183327,
+ "learning_rate": 0.002999710471471218,
+ "loss": 7.079098701477051,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.012408486567437649,
+ "learning_rate": 0.0029996835534525246,
+ "loss": 7.0790114402771,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.012035196647047997,
+ "learning_rate": 0.0029996554392511094,
+ "loss": 7.058756351470947,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.00899985246360302,
+ "learning_rate": 0.0029996261288896225,
+ "loss": 7.0739240646362305,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.01072906143963337,
+ "learning_rate": 0.002999595622391679,
+ "loss": 7.085537910461426,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.01047342736274004,
+ "learning_rate": 0.0029995639197818577,
+ "loss": 7.060049057006836,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.02342618815600872,
+ "learning_rate": 0.0029995310210857006,
+ "loss": 7.039482116699219,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.011256859637796879,
+ "learning_rate": 0.0029994969263297134,
+ "loss": 7.040412902832031,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.008407019078731537,
+ "learning_rate": 0.0029994616355413666,
+ "loss": 7.0364274978637695,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.011006148532032967,
+ "learning_rate": 0.0029994251487490916,
+ "loss": 6.991468906402588,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.009103255346417427,
+ "learning_rate": 0.0029993874659822875,
+ "loss": 7.001059532165527,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.007875788025557995,
+ "learning_rate": 0.0029993485872713125,
+ "loss": 6.9974260330200195,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.007872611284255981,
+ "learning_rate": 0.0029993085126474913,
+ "loss": 6.980216979980469,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.010071957483887672,
+ "learning_rate": 0.0029992672421431115,
+ "loss": 6.97799015045166,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.01209618803113699,
+ "learning_rate": 0.0029992247757914237,
+ "loss": 6.97202205657959,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.014884858392179012,
+ "learning_rate": 0.0029991811136266425,
+ "loss": 6.976570129394531,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.014953463338315487,
+ "learning_rate": 0.0029991362556839453,
+ "loss": 6.971050262451172,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.008371063508093357,
+ "learning_rate": 0.002999090201999474,
+ "loss": 6.955556869506836,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.011926881037652493,
+ "learning_rate": 0.002999042952610332,
+ "loss": 6.943947792053223,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.013422893360257149,
+ "learning_rate": 0.002998994507554589,
+ "loss": 6.910022735595703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.009672224521636963,
+ "learning_rate": 0.002998944866871275,
+ "loss": 6.8994855880737305,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.012590581551194191,
+ "learning_rate": 0.0029988940306003853,
+ "loss": 6.88709020614624,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.945146083831787,
+ "eval_runtime": 51.5801,
+ "eval_samples_per_second": 47.344,
+ "eval_steps_per_second": 1.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.016554836183786392,
+ "learning_rate": 0.002998841998782877,
+ "loss": 6.916609764099121,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.011085660196840763,
+ "learning_rate": 0.0029987887714606723,
+ "loss": 6.884910583496094,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.008596168830990791,
+ "learning_rate": 0.0029987343486766547,
+ "loss": 6.900066375732422,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.014487302862107754,
+ "learning_rate": 0.0029986787304746715,
+ "loss": 6.8693342208862305,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.015636997297406197,
+ "learning_rate": 0.002998621916899534,
+ "loss": 6.903670787811279,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.013275462202727795,
+ "learning_rate": 0.0029985639079970153,
+ "loss": 6.889361381530762,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.009044744074344635,
+ "learning_rate": 0.002998504703813852,
+ "loss": 6.8549981117248535,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.011327696964144707,
+ "learning_rate": 0.002998444304397745,
+ "loss": 6.856151103973389,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.01310331467539072,
+ "learning_rate": 0.0029983827097973554,
+ "loss": 6.834944725036621,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.010003753937780857,
+ "learning_rate": 0.00299831992006231,
+ "loss": 6.825613975524902,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.009928115643560886,
+ "learning_rate": 0.0029982559352431967,
+ "loss": 6.828935146331787,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.012302150949835777,
+ "learning_rate": 0.0029981907553915674,
+ "loss": 6.821207046508789,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.020555168390274048,
+ "learning_rate": 0.002998124380559936,
+ "loss": 6.8184895515441895,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.011417638510465622,
+ "learning_rate": 0.0029980568108017797,
+ "loss": 6.788754463195801,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.010772978886961937,
+ "learning_rate": 0.002997988046171538,
+ "loss": 6.7861433029174805,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.012753859162330627,
+ "learning_rate": 0.002997918086724614,
+ "loss": 6.788776397705078,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.011327114887535572,
+ "learning_rate": 0.0029978469325173717,
+ "loss": 6.779160022735596,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01235753484070301,
+ "learning_rate": 0.002997774583607139,
+ "loss": 6.799554824829102,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.01417156495153904,
+ "learning_rate": 0.0029977010400522074,
+ "loss": 6.77957820892334,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.021321553736925125,
+ "learning_rate": 0.002997626301911828,
+ "loss": 6.785632133483887,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.03614472970366478,
+ "learning_rate": 0.002997550369246217,
+ "loss": 6.836414337158203,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03934052586555481,
+ "learning_rate": 0.002997473242116551,
+ "loss": 6.854659557342529,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.016240060329437256,
+ "learning_rate": 0.0029973949205849707,
+ "loss": 6.789690971374512,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.031164905056357384,
+ "learning_rate": 0.0029973154047145782,
+ "loss": 6.8363776206970215,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.013051844201982021,
+ "learning_rate": 0.002997234694569438,
+ "loss": 6.803576469421387,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.022611448541283607,
+ "learning_rate": 0.002997152790214577,
+ "loss": 6.762107849121094,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.017812907695770264,
+ "learning_rate": 0.0029970696917159833,
+ "loss": 6.776218414306641,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.015453332103788853,
+ "learning_rate": 0.0029969853991406088,
+ "loss": 6.742618083953857,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.020630091428756714,
+ "learning_rate": 0.0029968999125563655,
+ "loss": 6.765475749969482,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.014672341756522655,
+ "learning_rate": 0.0029968132320321297,
+ "loss": 6.754009246826172,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.016540220007300377,
+ "learning_rate": 0.0029967253576377376,
+ "loss": 6.734752655029297,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.013495649211108685,
+ "learning_rate": 0.0029966362894439874,
+ "loss": 6.752676010131836,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015936989337205887,
+ "learning_rate": 0.0029965460275226405,
+ "loss": 6.725708961486816,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010431243106722832,
+ "learning_rate": 0.0029964545719464194,
+ "loss": 6.698376655578613,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01669321395456791,
+ "learning_rate": 0.0029963619227890083,
+ "loss": 6.711770534515381,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.013308997265994549,
+ "learning_rate": 0.0029962680801250514,
+ "loss": 6.705286979675293,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011648476123809814,
+ "learning_rate": 0.002996173044030159,
+ "loss": 6.690052509307861,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.03942161798477173,
+ "learning_rate": 0.0029960768145808965,
+ "loss": 6.711885452270508,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.01571274921298027,
+ "learning_rate": 0.002995979391854797,
+ "loss": 6.715847969055176,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.00829714722931385,
+ "learning_rate": 0.0029958807759303515,
+ "loss": 6.685308933258057,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.01374727487564087,
+ "learning_rate": 0.0029957809668870114,
+ "loss": 6.695744514465332,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.010857059620320797,
+ "learning_rate": 0.0029956799648051934,
+ "loss": 6.658294677734375,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.008897246792912483,
+ "learning_rate": 0.002995577769766272,
+ "loss": 6.642439842224121,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015880772843956947,
+ "learning_rate": 0.002995474381852584,
+ "loss": 6.674907684326172,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.021623898297548294,
+ "learning_rate": 0.0029953698011474265,
+ "loss": 6.66236686706543,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.03084627166390419,
+ "learning_rate": 0.0029952640277350596,
+ "loss": 6.666715621948242,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.040561504662036896,
+ "learning_rate": 0.002995157061700702,
+ "loss": 6.649062156677246,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.031126515939831734,
+ "learning_rate": 0.0029950489031305345,
+ "loss": 6.640542030334473,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.00872360821813345,
+ "learning_rate": 0.0029949395521116987,
+ "loss": 6.635221481323242,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.023432321846485138,
+ "learning_rate": 0.0029948290087322966,
+ "loss": 6.642196178436279,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.021096013486385345,
+ "learning_rate": 0.0029947172730813913,
+ "loss": 6.643655300140381,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.014494858682155609,
+ "learning_rate": 0.002994604345249006,
+ "loss": 6.612720966339111,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.05225599184632301,
+ "learning_rate": 0.0029944902253261248,
+ "loss": 6.60659646987915,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.014311269856989384,
+ "learning_rate": 0.0029943749134046915,
+ "loss": 6.54373025894165,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.010777472518384457,
+ "learning_rate": 0.0029942584095776115,
+ "loss": 6.599837303161621,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.012498079799115658,
+ "learning_rate": 0.002994140713938749,
+ "loss": 6.589657783508301,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.015775132924318314,
+ "learning_rate": 0.00299402182658293,
+ "loss": 6.577511787414551,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.017978230491280556,
+ "learning_rate": 0.0029939017476059396,
+ "loss": 6.58717155456543,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.026807745918631554,
+ "learning_rate": 0.0029937804771045236,
+ "loss": 6.578145503997803,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.04853282868862152,
+ "learning_rate": 0.0029936580151763872,
+ "loss": 6.609755516052246,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.1057916209101677,
+ "learning_rate": 0.002993534361920195,
+ "loss": 6.870367050170898,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.04925911873579025,
+ "learning_rate": 0.002993409517435574,
+ "loss": 6.871556282043457,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.05288556218147278,
+ "learning_rate": 0.0029932834818231077,
+ "loss": 6.808315753936768,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.03831710293889046,
+ "learning_rate": 0.0029931562551843405,
+ "loss": 6.827617645263672,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.03665117174386978,
+ "learning_rate": 0.002993027837621778,
+ "loss": 6.767284870147705,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.026216736063361168,
+ "learning_rate": 0.0029928982292388826,
+ "loss": 6.733659744262695,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.024796854704618454,
+ "learning_rate": 0.002992767430140078,
+ "loss": 6.759577751159668,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.02410992607474327,
+ "learning_rate": 0.002992635440430747,
+ "loss": 6.701825141906738,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.019924134016036987,
+ "learning_rate": 0.0029925022602172308,
+ "loss": 6.677121162414551,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.022516867145895958,
+ "learning_rate": 0.00299236788960683,
+ "loss": 6.668622970581055,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.025068556889891624,
+ "learning_rate": 0.002992232328707806,
+ "loss": 6.67718505859375,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.02574208751320839,
+ "learning_rate": 0.002992095577629376,
+ "loss": 6.679594039916992,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.657555103302002,
+ "eval_runtime": 51.0199,
+ "eval_samples_per_second": 47.864,
+ "eval_steps_per_second": 1.509,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.017298126593232155,
+ "learning_rate": 0.0029919576364817184,
+ "loss": 6.620774269104004,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.018840325996279716,
+ "learning_rate": 0.002991818505375971,
+ "loss": 6.613316535949707,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.020813113078475,
+ "learning_rate": 0.002991678184424228,
+ "loss": 6.621555328369141,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.026581479236483574,
+ "learning_rate": 0.0029915366737395437,
+ "loss": 6.613471031188965,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.02084030583500862,
+ "learning_rate": 0.0029913939734359306,
+ "loss": 6.559704780578613,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.01327502354979515,
+ "learning_rate": 0.00299125008362836,
+ "loss": 6.564910888671875,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.015091734007000923,
+ "learning_rate": 0.0029911050044327614,
+ "loss": 6.571262836456299,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.015463982708752155,
+ "learning_rate": 0.002990958735966023,
+ "loss": 6.565750598907471,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.013636826537549496,
+ "learning_rate": 0.0029908112783459894,
+ "loss": 6.57981014251709,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.010613273829221725,
+ "learning_rate": 0.0029906626316914655,
+ "loss": 6.538505554199219,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.01804237812757492,
+ "learning_rate": 0.0029905127961222127,
+ "loss": 6.5282511711120605,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.015131509862840176,
+ "learning_rate": 0.0029903617717589506,
+ "loss": 6.531335830688477,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.010518206283450127,
+ "learning_rate": 0.002990209558723358,
+ "loss": 6.521886825561523,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.011360022239387035,
+ "learning_rate": 0.002990056157138069,
+ "loss": 6.528301239013672,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.011556181125342846,
+ "learning_rate": 0.0029899015671266773,
+ "loss": 6.497917652130127,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.01365791168063879,
+ "learning_rate": 0.0029897457888137325,
+ "loss": 6.52849006652832,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.01091043371707201,
+ "learning_rate": 0.0029895888223247426,
+ "loss": 6.492197036743164,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.008484847843647003,
+ "learning_rate": 0.002989430667786173,
+ "loss": 6.497377395629883,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.00836879201233387,
+ "learning_rate": 0.0029892713253254466,
+ "loss": 6.495189666748047,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.009781497530639172,
+ "learning_rate": 0.0029891107950709406,
+ "loss": 6.454085826873779,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.009603648446500301,
+ "learning_rate": 0.0029889490771519934,
+ "loss": 6.499475479125977,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.007991855964064598,
+ "learning_rate": 0.0029887861716988975,
+ "loss": 6.460785865783691,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.008447187952697277,
+ "learning_rate": 0.0029886220788429025,
+ "loss": 6.504582405090332,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.010945498012006283,
+ "learning_rate": 0.0029884567987162156,
+ "loss": 6.49589729309082,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.011096215806901455,
+ "learning_rate": 0.0029882903314519997,
+ "loss": 6.460657119750977,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.013094739057123661,
+ "learning_rate": 0.0029881226771843747,
+ "loss": 6.462450981140137,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.015423037111759186,
+ "learning_rate": 0.0029879538360484165,
+ "loss": 6.45565128326416,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.0182222630828619,
+ "learning_rate": 0.002987783808180157,
+ "loss": 6.452227592468262,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.01981227844953537,
+ "learning_rate": 0.002987612593716585,
+ "loss": 6.477701663970947,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.026733700186014175,
+ "learning_rate": 0.002987440192795645,
+ "loss": 6.452212333679199,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.03504155948758125,
+ "learning_rate": 0.0029872666055562372,
+ "loss": 6.509665489196777,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.027970125898718834,
+ "learning_rate": 0.002987091832138217,
+ "loss": 6.487089157104492,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.013714591972529888,
+ "learning_rate": 0.0029869158726823974,
+ "loss": 6.469569206237793,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.016472695395350456,
+ "learning_rate": 0.0029867387273305453,
+ "loss": 6.461688995361328,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.018457358703017235,
+ "learning_rate": 0.0029865603962253833,
+ "loss": 6.490985870361328,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.011045100167393684,
+ "learning_rate": 0.0029863808795105896,
+ "loss": 6.403701305389404,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.014126897789537907,
+ "learning_rate": 0.0029862001773307972,
+ "loss": 6.379804611206055,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.014015335589647293,
+ "learning_rate": 0.0029860182898315954,
+ "loss": 6.404377460479736,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.010032658465206623,
+ "learning_rate": 0.002985835217159527,
+ "loss": 6.40291690826416,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.015469838865101337,
+ "learning_rate": 0.0029856509594620912,
+ "loss": 6.4414520263671875,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.011106910184025764,
+ "learning_rate": 0.0029854655168877397,
+ "loss": 6.396444320678711,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.01262320950627327,
+ "learning_rate": 0.0029852788895858816,
+ "loss": 6.385222434997559,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.012783976271748543,
+ "learning_rate": 0.002985091077706878,
+ "loss": 6.3818135261535645,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.01224067248404026,
+ "learning_rate": 0.0029849020814020468,
+ "loss": 6.37886905670166,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.010540086776018143,
+ "learning_rate": 0.0029847119008236576,
+ "loss": 6.37799072265625,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.012479913420975208,
+ "learning_rate": 0.0029845205361249366,
+ "loss": 6.365094184875488,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.010080077685415745,
+ "learning_rate": 0.0029843279874600622,
+ "loss": 6.371852874755859,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.013266298919916153,
+ "learning_rate": 0.002984134254984168,
+ "loss": 6.380039215087891,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.01014771405607462,
+ "learning_rate": 0.002983939338853339,
+ "loss": 6.370337963104248,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.008164579048752785,
+ "learning_rate": 0.0029837432392246185,
+ "loss": 6.331899642944336,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.01020442508161068,
+ "learning_rate": 0.002983545956255998,
+ "loss": 6.35077428817749,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.009898579679429531,
+ "learning_rate": 0.002983347490106426,
+ "loss": 6.371985912322998,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.009668919257819653,
+ "learning_rate": 0.002983147840935802,
+ "loss": 6.356776237487793,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.012560366652905941,
+ "learning_rate": 0.0029829470089049815,
+ "loss": 6.357911109924316,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.018435249105095863,
+ "learning_rate": 0.0029827449941757698,
+ "loss": 6.366821765899658,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.029046855866909027,
+ "learning_rate": 0.002982541796910927,
+ "loss": 6.367595672607422,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.043010637164115906,
+ "learning_rate": 0.002982337417274165,
+ "loss": 6.413588523864746,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.045470818877220154,
+ "learning_rate": 0.0029821318554301497,
+ "loss": 6.435015678405762,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.02148726023733616,
+ "learning_rate": 0.002981925111544498,
+ "loss": 6.330894470214844,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.026808688417077065,
+ "learning_rate": 0.00298171718578378,
+ "loss": 6.384839057922363,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.018495213240385056,
+ "learning_rate": 0.002981508078315517,
+ "loss": 6.36277961730957,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.025393597781658173,
+ "learning_rate": 0.0029812977893081834,
+ "loss": 6.3702239990234375,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020856289193034172,
+ "learning_rate": 0.002981086318931206,
+ "loss": 6.354547500610352,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.017827816307544708,
+ "learning_rate": 0.002980873667354962,
+ "loss": 6.339264869689941,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.02639714442193508,
+ "learning_rate": 0.002980659834750781,
+ "loss": 6.3254194259643555,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.015238096937537193,
+ "learning_rate": 0.002980444821290944,
+ "loss": 6.365842819213867,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.012863729149103165,
+ "learning_rate": 0.002980228627148684,
+ "loss": 6.356776714324951,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.014739899896085262,
+ "learning_rate": 0.0029800112524981845,
+ "loss": 6.332481384277344,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.013582171872258186,
+ "learning_rate": 0.0029797926975145797,
+ "loss": 6.311968803405762,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.012704637832939625,
+ "learning_rate": 0.0029795729623739555,
+ "loss": 6.321654319763184,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.012550560757517815,
+ "learning_rate": 0.0029793520472533495,
+ "loss": 6.298409461975098,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.011170165613293648,
+ "learning_rate": 0.002979129952330748,
+ "loss": 6.294830799102783,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.326757907867432,
+ "eval_runtime": 51.6566,
+ "eval_samples_per_second": 47.274,
+ "eval_steps_per_second": 1.491,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.010947446338832378,
+ "learning_rate": 0.0029789066777850897,
+ "loss": 6.286581039428711,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.010097995400428772,
+ "learning_rate": 0.002978682223796262,
+ "loss": 6.288301467895508,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.01224475260823965,
+ "learning_rate": 0.002978456590545104,
+ "loss": 6.264986991882324,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.009734537452459335,
+ "learning_rate": 0.0029782297782134045,
+ "loss": 6.304243087768555,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.01136956550180912,
+ "learning_rate": 0.0029780017869839013,
+ "loss": 6.244938373565674,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.00935293547809124,
+ "learning_rate": 0.002977772617040284,
+ "loss": 6.251852035522461,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.011118784546852112,
+ "learning_rate": 0.0029775422685671894,
+ "loss": 6.256633758544922,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.010837923735380173,
+ "learning_rate": 0.0029773107417502055,
+ "loss": 6.263411045074463,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.011920305900275707,
+ "learning_rate": 0.0029770780367758703,
+ "loss": 6.239311218261719,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.014117483049631119,
+ "learning_rate": 0.0029768441538316685,
+ "loss": 6.253314018249512,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.022502148523926735,
+ "learning_rate": 0.002976609093106036,
+ "loss": 6.246156692504883,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.03348720073699951,
+ "learning_rate": 0.002976372854788357,
+ "loss": 6.255914688110352,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.03427121415734291,
+ "learning_rate": 0.002976135439068965,
+ "loss": 6.289946556091309,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.02407355234026909,
+ "learning_rate": 0.002975896846139141,
+ "loss": 6.288534164428711,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.017861567437648773,
+ "learning_rate": 0.0029756570761911147,
+ "loss": 6.258049964904785,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.019980227574706078,
+ "learning_rate": 0.0029754161294180646,
+ "loss": 6.250895023345947,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.01443127728998661,
+ "learning_rate": 0.0029751740060141177,
+ "loss": 6.212200164794922,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.014975055120885372,
+ "learning_rate": 0.0029749307061743475,
+ "loss": 6.230568885803223,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01408296450972557,
+ "learning_rate": 0.0029746862300947773,
+ "loss": 6.243950366973877,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.011649327352643013,
+ "learning_rate": 0.0029744405779723766,
+ "loss": 6.197744846343994,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.013392188586294651,
+ "learning_rate": 0.002974193750005063,
+ "loss": 6.207277297973633,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.010902997106313705,
+ "learning_rate": 0.002973945746391701,
+ "loss": 6.166373252868652,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.012915392406284809,
+ "learning_rate": 0.0029736965673321032,
+ "loss": 6.186316013336182,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.010772830806672573,
+ "learning_rate": 0.0029734462130270273,
+ "loss": 6.172578811645508,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.010274911299347878,
+ "learning_rate": 0.002973194683678181,
+ "loss": 6.206851005554199,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.013210494071245193,
+ "learning_rate": 0.0029729419794882164,
+ "loss": 6.21115779876709,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.013372313231229782,
+ "learning_rate": 0.0029726881006607326,
+ "loss": 6.18442440032959,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.008105908520519733,
+ "learning_rate": 0.0029724330474002743,
+ "loss": 6.204354286193848,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.014167108573019505,
+ "learning_rate": 0.0029721768199123346,
+ "loss": 6.180662155151367,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.008339803665876389,
+ "learning_rate": 0.002971919418403351,
+ "loss": 6.150079727172852,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.015939386561512947,
+ "learning_rate": 0.0029716608430807073,
+ "loss": 6.175203323364258,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.010777109302580357,
+ "learning_rate": 0.0029714010941527327,
+ "loss": 6.166106224060059,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.011274735443294048,
+ "learning_rate": 0.002971140171828702,
+ "loss": 6.163111686706543,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013626808300614357,
+ "learning_rate": 0.002970878076318836,
+ "loss": 6.140850067138672,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.016907762736082077,
+ "learning_rate": 0.0029706148078343,
+ "loss": 6.192012786865234,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.018817441537976265,
+ "learning_rate": 0.002970350366587205,
+ "loss": 6.162238121032715,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.02829768694937229,
+ "learning_rate": 0.0029700847527906068,
+ "loss": 6.159066200256348,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.02034052088856697,
+ "learning_rate": 0.002969817966658505,
+ "loss": 6.164158344268799,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.01816369965672493,
+ "learning_rate": 0.0029695500084058447,
+ "loss": 6.132316589355469,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.013685261830687523,
+ "learning_rate": 0.0029692808782485146,
+ "loss": 6.152181625366211,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.009791149757802486,
+ "learning_rate": 0.0029690105764033485,
+ "loss": 6.138544082641602,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.014827069826424122,
+ "learning_rate": 0.002968739103088124,
+ "loss": 6.135766983032227,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.01391249243170023,
+ "learning_rate": 0.0029684664585215618,
+ "loss": 6.146172523498535,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.012729899026453495,
+ "learning_rate": 0.002968192642923326,
+ "loss": 6.122613430023193,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.014175458811223507,
+ "learning_rate": 0.002967917656514026,
+ "loss": 6.12574577331543,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.010958840139210224,
+ "learning_rate": 0.002967641499515213,
+ "loss": 6.111088752746582,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.015380816534161568,
+ "learning_rate": 0.0029673641721493817,
+ "loss": 6.100615501403809,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.025339096784591675,
+ "learning_rate": 0.00296708567463997,
+ "loss": 6.139659881591797,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.03695995733141899,
+ "learning_rate": 0.002966806007211358,
+ "loss": 6.160996437072754,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.0408046580851078,
+ "learning_rate": 0.0029665251700888686,
+ "loss": 6.173980712890625,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.025022024288773537,
+ "learning_rate": 0.0029662431634987674,
+ "loss": 6.145794868469238,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.027254154905676842,
+ "learning_rate": 0.002965959987668262,
+ "loss": 6.16928243637085,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.06191903352737427,
+ "learning_rate": 0.0029656756428255017,
+ "loss": 6.138525009155273,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0348956398665905,
+ "learning_rate": 0.002965390129199579,
+ "loss": 6.185477256774902,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.020859677344560623,
+ "learning_rate": 0.002965103447020526,
+ "loss": 6.148846626281738,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.022810814902186394,
+ "learning_rate": 0.0029648155965193174,
+ "loss": 6.1701178550720215,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.017981817945837975,
+ "learning_rate": 0.00296452657792787,
+ "loss": 6.1323771476745605,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.014905866235494614,
+ "learning_rate": 0.00296423639147904,
+ "loss": 6.123426914215088,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.01463468000292778,
+ "learning_rate": 0.002963945037406626,
+ "loss": 6.146122932434082,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.021616583690047264,
+ "learning_rate": 0.002963652515945366,
+ "loss": 6.093459606170654,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.01550968922674656,
+ "learning_rate": 0.00296335882733094,
+ "loss": 6.114929676055908,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.014597428031265736,
+ "learning_rate": 0.0029630639717999675,
+ "loss": 6.113945960998535,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.018249280750751495,
+ "learning_rate": 0.002962767949590008,
+ "loss": 6.130507469177246,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.014160148799419403,
+ "learning_rate": 0.0029624707609395614,
+ "loss": 6.086946487426758,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.015563832595944405,
+ "learning_rate": 0.002962172406088067,
+ "loss": 6.083332538604736,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.012240015901625156,
+ "learning_rate": 0.002961872885275904,
+ "loss": 6.0822906494140625,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.0141651201993227,
+ "learning_rate": 0.002961572198744391,
+ "loss": 6.049562454223633,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.010004645213484764,
+ "learning_rate": 0.0029612703467357866,
+ "loss": 6.067063808441162,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.011790835298597813,
+ "learning_rate": 0.0029609673294932856,
+ "loss": 6.062193870544434,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.011088060215115547,
+ "learning_rate": 0.002960663147261025,
+ "loss": 6.070950031280518,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.010387620888650417,
+ "learning_rate": 0.002960357800284078,
+ "loss": 6.032423973083496,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.012074458412826061,
+ "learning_rate": 0.002960051288808458,
+ "loss": 6.066788673400879,
+ "step": 360
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.534806680666112e+17,
+ "train_batch_size": 16,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/baseline/checkpoint-360/training_args.bin b/runs/baseline/checkpoint-360/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7
--- /dev/null
+++ b/runs/baseline/checkpoint-360/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97
+size 5265
diff --git a/runs/baseline/checkpoint-3600/chat_template.jinja b/runs/baseline/checkpoint-3600/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/baseline/checkpoint-3600/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/baseline/checkpoint-3600/config.json b/runs/baseline/checkpoint-3600/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/baseline/checkpoint-3600/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/baseline/checkpoint-3600/generation_config.json b/runs/baseline/checkpoint-3600/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/baseline/checkpoint-3600/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/baseline/checkpoint-3600/model.safetensors b/runs/baseline/checkpoint-3600/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..3f5e82e28e3d58b9343dd1945b5a105f60539327
--- /dev/null
+++ b/runs/baseline/checkpoint-3600/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:72ce76048e8d4b7c3387c7d5b90f6274666606ec81c5f33ae4591fc0e12177dc
+size 583356232
diff --git a/runs/baseline/checkpoint-3600/optimizer.pt b/runs/baseline/checkpoint-3600/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..f229dcb5a60c628efe4c5e718334ddd1f2a29801
--- /dev/null
+++ b/runs/baseline/checkpoint-3600/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4fe36266830f9e17f5f5a3ffc66f118b6e2bed37f92d7ff3bf9956544a04fc42
+size 1166825803
diff --git a/runs/baseline/checkpoint-3600/rng_state_0.pth b/runs/baseline/checkpoint-3600/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..42e5fb760a4884c24f3899e0e2c34f9ab7d35d5e
--- /dev/null
+++ b/runs/baseline/checkpoint-3600/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b9abac844de24b211d8b88d44e1866f089edee417850520a358afe34f866a821
+size 14917
diff --git a/runs/baseline/checkpoint-3600/rng_state_1.pth b/runs/baseline/checkpoint-3600/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..d6f10c7be65a25152763463abf33eb700d784288
--- /dev/null
+++ b/runs/baseline/checkpoint-3600/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:68ba205d24a50bf8cdf4686014dfa7b5557eb04aefa63431dced1c3d75cbacab
+size 14917
diff --git a/runs/baseline/checkpoint-3600/scheduler.pt b/runs/baseline/checkpoint-3600/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..f967f3272dea4dcd87d1b7218164c5b8c405a0a2
--- /dev/null
+++ b/runs/baseline/checkpoint-3600/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:96be326da54197356c9c3f871e031619329af2df05a743623e3ef304de77920c
+size 1465
diff --git a/runs/baseline/checkpoint-3600/tokenizer.json b/runs/baseline/checkpoint-3600/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/baseline/checkpoint-3600/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/baseline/checkpoint-3600/tokenizer_config.json b/runs/baseline/checkpoint-3600/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/baseline/checkpoint-3600/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/baseline/checkpoint-3600/trainer_state.json b/runs/baseline/checkpoint-3600/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..c80990d5a85c7660d3d9ce92b20c3d5aa2c6ea4a
--- /dev/null
+++ b/runs/baseline/checkpoint-3600/trainer_state.json
@@ -0,0 +1,25626 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 50.0,
+ "eval_steps": 500,
+ "global_step": 3600,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.12165174633264542,
+ "learning_rate": 0.0,
+ "loss": 12.027063369750977,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.12109383195638657,
+ "learning_rate": 3e-05,
+ "loss": 12.028635025024414,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.12690064311027527,
+ "learning_rate": 6e-05,
+ "loss": 11.894524574279785,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.1537286937236786,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.672262191772461,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.11778244376182556,
+ "learning_rate": 0.00012,
+ "loss": 11.433671951293945,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.10874935984611511,
+ "learning_rate": 0.00015000000000000001,
+ "loss": 11.270309448242188,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.10641921311616898,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 11.155481338500977,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.1048506498336792,
+ "learning_rate": 0.00021000000000000004,
+ "loss": 11.066675186157227,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.10546834021806717,
+ "learning_rate": 0.00024,
+ "loss": 10.952295303344727,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.10614268481731415,
+ "learning_rate": 0.00027,
+ "loss": 10.811386108398438,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10529010742902756,
+ "learning_rate": 0.00030000000000000003,
+ "loss": 10.657712936401367,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.1051444336771965,
+ "learning_rate": 0.00033,
+ "loss": 10.47591781616211,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10420846194028854,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 10.280141830444336,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10323625057935715,
+ "learning_rate": 0.00039000000000000005,
+ "loss": 10.076366424560547,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10255401581525803,
+ "learning_rate": 0.00042000000000000007,
+ "loss": 9.856939315795898,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10151027888059616,
+ "learning_rate": 0.00045,
+ "loss": 9.632134437561035,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.1006326898932457,
+ "learning_rate": 0.00048,
+ "loss": 9.401637077331543,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.09820562601089478,
+ "learning_rate": 0.00051,
+ "loss": 9.177400588989258,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.0956626906991005,
+ "learning_rate": 0.00054,
+ "loss": 8.95518684387207,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.0922091156244278,
+ "learning_rate": 0.00057,
+ "loss": 8.75066089630127,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.08593932539224625,
+ "learning_rate": 0.0006000000000000001,
+ "loss": 8.557976722717285,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.08014766871929169,
+ "learning_rate": 0.00063,
+ "loss": 8.360604286193848,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.07381244748830795,
+ "learning_rate": 0.00066,
+ "loss": 8.210375785827637,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.06455052644014359,
+ "learning_rate": 0.0006900000000000001,
+ "loss": 8.07168960571289,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.054115474224090576,
+ "learning_rate": 0.0007199999999999999,
+ "loss": 7.960069179534912,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.043575264513492584,
+ "learning_rate": 0.00075,
+ "loss": 7.850685119628906,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.02976679615676403,
+ "learning_rate": 0.0007800000000000001,
+ "loss": 7.824377059936523,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.020552165806293488,
+ "learning_rate": 0.0008100000000000001,
+ "loss": 7.7695417404174805,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.018530743196606636,
+ "learning_rate": 0.0008400000000000001,
+ "loss": 7.782247543334961,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.02716820128262043,
+ "learning_rate": 0.00087,
+ "loss": 7.754945278167725,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.03638409823179245,
+ "learning_rate": 0.0009,
+ "loss": 7.799381256103516,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.041579652577638626,
+ "learning_rate": 0.00093,
+ "loss": 7.776629447937012,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.0304960235953331,
+ "learning_rate": 0.00096,
+ "loss": 7.749439239501953,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.053728822618722916,
+ "learning_rate": 0.00099,
+ "loss": 7.745059490203857,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.036322493106126785,
+ "learning_rate": 0.00102,
+ "loss": 7.752074718475342,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.03232959285378456,
+ "learning_rate": 0.00105,
+ "loss": 7.704257011413574,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.03246184438467026,
+ "learning_rate": 0.00108,
+ "loss": 7.667045593261719,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.036655545234680176,
+ "learning_rate": 0.00111,
+ "loss": 7.623445987701416,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.021042725071310997,
+ "learning_rate": 0.00114,
+ "loss": 7.582821846008301,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.022982975468039513,
+ "learning_rate": 0.00117,
+ "loss": 7.58036994934082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.024781864136457443,
+ "learning_rate": 0.0012000000000000001,
+ "loss": 7.5903000831604,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.03667089343070984,
+ "learning_rate": 0.00123,
+ "loss": 7.588233947753906,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.04083017632365227,
+ "learning_rate": 0.00126,
+ "loss": 7.548343658447266,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.02585385926067829,
+ "learning_rate": 0.00129,
+ "loss": 7.604833602905273,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.018793931230902672,
+ "learning_rate": 0.00132,
+ "loss": 7.56705379486084,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.025538936257362366,
+ "learning_rate": 0.00135,
+ "loss": 7.562107086181641,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.02819318138062954,
+ "learning_rate": 0.0013800000000000002,
+ "loss": 7.517213821411133,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.016239255666732788,
+ "learning_rate": 0.00141,
+ "loss": 7.4780755043029785,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.024470701813697815,
+ "learning_rate": 0.0014399999999999999,
+ "loss": 7.505887985229492,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.015499847941100597,
+ "learning_rate": 0.00147,
+ "loss": 7.484469890594482,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.017785200849175453,
+ "learning_rate": 0.0015,
+ "loss": 7.481822967529297,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.015952805057168007,
+ "learning_rate": 0.0015300000000000001,
+ "loss": 7.484804153442383,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.015771260485053062,
+ "learning_rate": 0.0015600000000000002,
+ "loss": 7.4495439529418945,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.024370530620217323,
+ "learning_rate": 0.00159,
+ "loss": 7.471208572387695,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.018173737451434135,
+ "learning_rate": 0.0016200000000000001,
+ "loss": 7.438447952270508,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.020564094185829163,
+ "learning_rate": 0.0016500000000000002,
+ "loss": 7.437922954559326,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.016154944896697998,
+ "learning_rate": 0.0016800000000000003,
+ "loss": 7.429378032684326,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.016817007213830948,
+ "learning_rate": 0.00171,
+ "loss": 7.406428337097168,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.03460947796702385,
+ "learning_rate": 0.00174,
+ "loss": 7.393331527709961,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.15410488843917847,
+ "learning_rate": 0.0017699999999999999,
+ "loss": 7.645856857299805,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.07443148642778397,
+ "learning_rate": 0.0018,
+ "loss": 7.549508094787598,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.0507572740316391,
+ "learning_rate": 0.00183,
+ "loss": 7.517921447753906,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.09682746231555939,
+ "learning_rate": 0.00186,
+ "loss": 7.543978691101074,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.2571517527103424,
+ "learning_rate": 0.00189,
+ "loss": 7.768949508666992,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.0317784883081913,
+ "learning_rate": 0.00192,
+ "loss": 7.509063720703125,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.0353970006108284,
+ "learning_rate": 0.0019500000000000001,
+ "loss": 7.522322654724121,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.03663994371891022,
+ "learning_rate": 0.00198,
+ "loss": 7.504723072052002,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.025481393560767174,
+ "learning_rate": 0.00201,
+ "loss": 7.452597141265869,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.03099915198981762,
+ "learning_rate": 0.00204,
+ "loss": 7.455441474914551,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.022921698167920113,
+ "learning_rate": 0.00207,
+ "loss": 7.4454827308654785,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.032388731837272644,
+ "learning_rate": 0.0021,
+ "loss": 7.467103958129883,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.022428609430789948,
+ "learning_rate": 0.00213,
+ "loss": 7.467540264129639,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.469719409942627,
+ "eval_runtime": 50.9714,
+ "eval_samples_per_second": 47.909,
+ "eval_steps_per_second": 1.511,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.017480911687016487,
+ "learning_rate": 0.00216,
+ "loss": 7.455810546875,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.016675131395459175,
+ "learning_rate": 0.00219,
+ "loss": 7.474848747253418,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.017450228333473206,
+ "learning_rate": 0.00222,
+ "loss": 7.442611217498779,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.016823675483465195,
+ "learning_rate": 0.0022500000000000003,
+ "loss": 7.454628944396973,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.016717543825507164,
+ "learning_rate": 0.00228,
+ "loss": 7.425317764282227,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.01644429750740528,
+ "learning_rate": 0.00231,
+ "loss": 7.4102864265441895,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.013104400597512722,
+ "learning_rate": 0.00234,
+ "loss": 7.4559173583984375,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.01175684668123722,
+ "learning_rate": 0.00237,
+ "loss": 7.438898086547852,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.013736684806644917,
+ "learning_rate": 0.0024000000000000002,
+ "loss": 7.449005603790283,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01176756713539362,
+ "learning_rate": 0.0024300000000000003,
+ "loss": 7.428421974182129,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.011860918253660202,
+ "learning_rate": 0.00246,
+ "loss": 7.411491394042969,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.010093477554619312,
+ "learning_rate": 0.00249,
+ "loss": 7.39311408996582,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.009453288279473782,
+ "learning_rate": 0.00252,
+ "loss": 7.387656211853027,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.010311692021787167,
+ "learning_rate": 0.00255,
+ "loss": 7.401076316833496,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.010053195990622044,
+ "learning_rate": 0.00258,
+ "loss": 7.39655876159668,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.009758037514984608,
+ "learning_rate": 0.00261,
+ "loss": 7.423122882843018,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008394946344196796,
+ "learning_rate": 0.00264,
+ "loss": 7.383601188659668,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.008457187563180923,
+ "learning_rate": 0.00267,
+ "loss": 7.400934219360352,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.010601334273815155,
+ "learning_rate": 0.0027,
+ "loss": 7.400241851806641,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.009009373374283314,
+ "learning_rate": 0.0027300000000000002,
+ "loss": 7.377348899841309,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.00737508200109005,
+ "learning_rate": 0.0027600000000000003,
+ "loss": 7.385881423950195,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.0075711836107075214,
+ "learning_rate": 0.0027900000000000004,
+ "loss": 7.348570823669434,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.00846901722252369,
+ "learning_rate": 0.00282,
+ "loss": 7.364287853240967,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.007160635199397802,
+ "learning_rate": 0.00285,
+ "loss": 7.3881635665893555,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.00682067358866334,
+ "learning_rate": 0.0028799999999999997,
+ "loss": 7.348663330078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.009108101949095726,
+ "learning_rate": 0.00291,
+ "loss": 7.356302261352539,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.015828927978873253,
+ "learning_rate": 0.00294,
+ "loss": 7.347576141357422,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.009683096781373024,
+ "learning_rate": 0.00297,
+ "loss": 7.323967933654785,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.010833270847797394,
+ "learning_rate": 0.003,
+ "loss": 7.3139495849609375,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.014905400574207306,
+ "learning_rate": 0.002999999401781161,
+ "loss": 7.302030086517334,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.023674853146076202,
+ "learning_rate": 0.002999997607125127,
+ "loss": 7.311811923980713,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.050460800528526306,
+ "learning_rate": 0.002999994616033342,
+ "loss": 7.41972541809082,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.02193174883723259,
+ "learning_rate": 0.002999990428508218,
+ "loss": 7.294468879699707,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.036261629313230515,
+ "learning_rate": 0.0029999850445531283,
+ "loss": 7.295568466186523,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.02734929881989956,
+ "learning_rate": 0.002999978464172409,
+ "loss": 7.309239864349365,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.017884107306599617,
+ "learning_rate": 0.002999970687371363,
+ "loss": 7.2427592277526855,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01700308918952942,
+ "learning_rate": 0.0029999617141562573,
+ "loss": 7.241730213165283,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.0209935512393713,
+ "learning_rate": 0.0029999515445343194,
+ "loss": 7.262085437774658,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.023207105696201324,
+ "learning_rate": 0.0029999401785137437,
+ "loss": 7.25450325012207,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.01907353848218918,
+ "learning_rate": 0.002999927616103687,
+ "loss": 7.23149299621582,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.016149666160345078,
+ "learning_rate": 0.0029999138573142715,
+ "loss": 7.210232734680176,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.017513928934931755,
+ "learning_rate": 0.002999898902156582,
+ "loss": 7.218451976776123,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.015002547763288021,
+ "learning_rate": 0.002999882750642667,
+ "loss": 7.169316291809082,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.014233081601560116,
+ "learning_rate": 0.0029998654027855406,
+ "loss": 7.194825172424316,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.01283260714262724,
+ "learning_rate": 0.0029998468585991783,
+ "loss": 7.15669059753418,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.014416799880564213,
+ "learning_rate": 0.0029998271180985223,
+ "loss": 7.143369674682617,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.014337058179080486,
+ "learning_rate": 0.0029998061812994757,
+ "loss": 7.127734184265137,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.016406210139393806,
+ "learning_rate": 0.002999784048218908,
+ "loss": 7.086504936218262,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.01680421642959118,
+ "learning_rate": 0.002999760718874652,
+ "loss": 7.089420318603516,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.010665548034012318,
+ "learning_rate": 0.0029997361932855016,
+ "loss": 7.102463722229004,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013479137793183327,
+ "learning_rate": 0.002999710471471218,
+ "loss": 7.079098701477051,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.012408486567437649,
+ "learning_rate": 0.0029996835534525246,
+ "loss": 7.0790114402771,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.012035196647047997,
+ "learning_rate": 0.0029996554392511094,
+ "loss": 7.058756351470947,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.00899985246360302,
+ "learning_rate": 0.0029996261288896225,
+ "loss": 7.0739240646362305,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.01072906143963337,
+ "learning_rate": 0.002999595622391679,
+ "loss": 7.085537910461426,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.01047342736274004,
+ "learning_rate": 0.0029995639197818577,
+ "loss": 7.060049057006836,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.02342618815600872,
+ "learning_rate": 0.0029995310210857006,
+ "loss": 7.039482116699219,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.011256859637796879,
+ "learning_rate": 0.0029994969263297134,
+ "loss": 7.040412902832031,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.008407019078731537,
+ "learning_rate": 0.0029994616355413666,
+ "loss": 7.0364274978637695,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.011006148532032967,
+ "learning_rate": 0.0029994251487490916,
+ "loss": 6.991468906402588,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.009103255346417427,
+ "learning_rate": 0.0029993874659822875,
+ "loss": 7.001059532165527,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.007875788025557995,
+ "learning_rate": 0.0029993485872713125,
+ "loss": 6.9974260330200195,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.007872611284255981,
+ "learning_rate": 0.0029993085126474913,
+ "loss": 6.980216979980469,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.010071957483887672,
+ "learning_rate": 0.0029992672421431115,
+ "loss": 6.97799015045166,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.01209618803113699,
+ "learning_rate": 0.0029992247757914237,
+ "loss": 6.97202205657959,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.014884858392179012,
+ "learning_rate": 0.0029991811136266425,
+ "loss": 6.976570129394531,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.014953463338315487,
+ "learning_rate": 0.0029991362556839453,
+ "loss": 6.971050262451172,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.008371063508093357,
+ "learning_rate": 0.002999090201999474,
+ "loss": 6.955556869506836,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.011926881037652493,
+ "learning_rate": 0.002999042952610332,
+ "loss": 6.943947792053223,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.013422893360257149,
+ "learning_rate": 0.002998994507554589,
+ "loss": 6.910022735595703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.009672224521636963,
+ "learning_rate": 0.002998944866871275,
+ "loss": 6.8994855880737305,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.012590581551194191,
+ "learning_rate": 0.0029988940306003853,
+ "loss": 6.88709020614624,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.945146083831787,
+ "eval_runtime": 51.5801,
+ "eval_samples_per_second": 47.344,
+ "eval_steps_per_second": 1.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.016554836183786392,
+ "learning_rate": 0.002998841998782877,
+ "loss": 6.916609764099121,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.011085660196840763,
+ "learning_rate": 0.0029987887714606723,
+ "loss": 6.884910583496094,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.008596168830990791,
+ "learning_rate": 0.0029987343486766547,
+ "loss": 6.900066375732422,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.014487302862107754,
+ "learning_rate": 0.0029986787304746715,
+ "loss": 6.8693342208862305,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.015636997297406197,
+ "learning_rate": 0.002998621916899534,
+ "loss": 6.903670787811279,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.013275462202727795,
+ "learning_rate": 0.0029985639079970153,
+ "loss": 6.889361381530762,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.009044744074344635,
+ "learning_rate": 0.002998504703813852,
+ "loss": 6.8549981117248535,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.011327696964144707,
+ "learning_rate": 0.002998444304397745,
+ "loss": 6.856151103973389,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.01310331467539072,
+ "learning_rate": 0.0029983827097973554,
+ "loss": 6.834944725036621,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.010003753937780857,
+ "learning_rate": 0.00299831992006231,
+ "loss": 6.825613975524902,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.009928115643560886,
+ "learning_rate": 0.0029982559352431967,
+ "loss": 6.828935146331787,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.012302150949835777,
+ "learning_rate": 0.0029981907553915674,
+ "loss": 6.821207046508789,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.020555168390274048,
+ "learning_rate": 0.002998124380559936,
+ "loss": 6.8184895515441895,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.011417638510465622,
+ "learning_rate": 0.0029980568108017797,
+ "loss": 6.788754463195801,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.010772978886961937,
+ "learning_rate": 0.002997988046171538,
+ "loss": 6.7861433029174805,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.012753859162330627,
+ "learning_rate": 0.002997918086724614,
+ "loss": 6.788776397705078,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.011327114887535572,
+ "learning_rate": 0.0029978469325173717,
+ "loss": 6.779160022735596,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01235753484070301,
+ "learning_rate": 0.002997774583607139,
+ "loss": 6.799554824829102,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.01417156495153904,
+ "learning_rate": 0.0029977010400522074,
+ "loss": 6.77957820892334,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.021321553736925125,
+ "learning_rate": 0.002997626301911828,
+ "loss": 6.785632133483887,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.03614472970366478,
+ "learning_rate": 0.002997550369246217,
+ "loss": 6.836414337158203,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03934052586555481,
+ "learning_rate": 0.002997473242116551,
+ "loss": 6.854659557342529,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.016240060329437256,
+ "learning_rate": 0.0029973949205849707,
+ "loss": 6.789690971374512,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.031164905056357384,
+ "learning_rate": 0.0029973154047145782,
+ "loss": 6.8363776206970215,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.013051844201982021,
+ "learning_rate": 0.002997234694569438,
+ "loss": 6.803576469421387,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.022611448541283607,
+ "learning_rate": 0.002997152790214577,
+ "loss": 6.762107849121094,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.017812907695770264,
+ "learning_rate": 0.0029970696917159833,
+ "loss": 6.776218414306641,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.015453332103788853,
+ "learning_rate": 0.0029969853991406088,
+ "loss": 6.742618083953857,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.020630091428756714,
+ "learning_rate": 0.0029968999125563655,
+ "loss": 6.765475749969482,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.014672341756522655,
+ "learning_rate": 0.0029968132320321297,
+ "loss": 6.754009246826172,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.016540220007300377,
+ "learning_rate": 0.0029967253576377376,
+ "loss": 6.734752655029297,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.013495649211108685,
+ "learning_rate": 0.0029966362894439874,
+ "loss": 6.752676010131836,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015936989337205887,
+ "learning_rate": 0.0029965460275226405,
+ "loss": 6.725708961486816,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010431243106722832,
+ "learning_rate": 0.0029964545719464194,
+ "loss": 6.698376655578613,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01669321395456791,
+ "learning_rate": 0.0029963619227890083,
+ "loss": 6.711770534515381,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.013308997265994549,
+ "learning_rate": 0.0029962680801250514,
+ "loss": 6.705286979675293,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011648476123809814,
+ "learning_rate": 0.002996173044030159,
+ "loss": 6.690052509307861,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.03942161798477173,
+ "learning_rate": 0.0029960768145808965,
+ "loss": 6.711885452270508,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.01571274921298027,
+ "learning_rate": 0.002995979391854797,
+ "loss": 6.715847969055176,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.00829714722931385,
+ "learning_rate": 0.0029958807759303515,
+ "loss": 6.685308933258057,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.01374727487564087,
+ "learning_rate": 0.0029957809668870114,
+ "loss": 6.695744514465332,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.010857059620320797,
+ "learning_rate": 0.0029956799648051934,
+ "loss": 6.658294677734375,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.008897246792912483,
+ "learning_rate": 0.002995577769766272,
+ "loss": 6.642439842224121,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015880772843956947,
+ "learning_rate": 0.002995474381852584,
+ "loss": 6.674907684326172,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.021623898297548294,
+ "learning_rate": 0.0029953698011474265,
+ "loss": 6.66236686706543,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.03084627166390419,
+ "learning_rate": 0.0029952640277350596,
+ "loss": 6.666715621948242,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.040561504662036896,
+ "learning_rate": 0.002995157061700702,
+ "loss": 6.649062156677246,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.031126515939831734,
+ "learning_rate": 0.0029950489031305345,
+ "loss": 6.640542030334473,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.00872360821813345,
+ "learning_rate": 0.0029949395521116987,
+ "loss": 6.635221481323242,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.023432321846485138,
+ "learning_rate": 0.0029948290087322966,
+ "loss": 6.642196178436279,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.021096013486385345,
+ "learning_rate": 0.0029947172730813913,
+ "loss": 6.643655300140381,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.014494858682155609,
+ "learning_rate": 0.002994604345249006,
+ "loss": 6.612720966339111,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.05225599184632301,
+ "learning_rate": 0.0029944902253261248,
+ "loss": 6.60659646987915,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.014311269856989384,
+ "learning_rate": 0.0029943749134046915,
+ "loss": 6.54373025894165,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.010777472518384457,
+ "learning_rate": 0.0029942584095776115,
+ "loss": 6.599837303161621,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.012498079799115658,
+ "learning_rate": 0.002994140713938749,
+ "loss": 6.589657783508301,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.015775132924318314,
+ "learning_rate": 0.00299402182658293,
+ "loss": 6.577511787414551,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.017978230491280556,
+ "learning_rate": 0.0029939017476059396,
+ "loss": 6.58717155456543,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.026807745918631554,
+ "learning_rate": 0.0029937804771045236,
+ "loss": 6.578145503997803,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.04853282868862152,
+ "learning_rate": 0.0029936580151763872,
+ "loss": 6.609755516052246,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.1057916209101677,
+ "learning_rate": 0.002993534361920195,
+ "loss": 6.870367050170898,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.04925911873579025,
+ "learning_rate": 0.002993409517435574,
+ "loss": 6.871556282043457,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.05288556218147278,
+ "learning_rate": 0.0029932834818231077,
+ "loss": 6.808315753936768,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.03831710293889046,
+ "learning_rate": 0.0029931562551843405,
+ "loss": 6.827617645263672,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.03665117174386978,
+ "learning_rate": 0.002993027837621778,
+ "loss": 6.767284870147705,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.026216736063361168,
+ "learning_rate": 0.0029928982292388826,
+ "loss": 6.733659744262695,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.024796854704618454,
+ "learning_rate": 0.002992767430140078,
+ "loss": 6.759577751159668,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.02410992607474327,
+ "learning_rate": 0.002992635440430747,
+ "loss": 6.701825141906738,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.019924134016036987,
+ "learning_rate": 0.0029925022602172308,
+ "loss": 6.677121162414551,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.022516867145895958,
+ "learning_rate": 0.00299236788960683,
+ "loss": 6.668622970581055,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.025068556889891624,
+ "learning_rate": 0.002992232328707806,
+ "loss": 6.67718505859375,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.02574208751320839,
+ "learning_rate": 0.002992095577629376,
+ "loss": 6.679594039916992,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.657555103302002,
+ "eval_runtime": 51.0199,
+ "eval_samples_per_second": 47.864,
+ "eval_steps_per_second": 1.509,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.017298126593232155,
+ "learning_rate": 0.0029919576364817184,
+ "loss": 6.620774269104004,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.018840325996279716,
+ "learning_rate": 0.002991818505375971,
+ "loss": 6.613316535949707,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.020813113078475,
+ "learning_rate": 0.002991678184424228,
+ "loss": 6.621555328369141,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.026581479236483574,
+ "learning_rate": 0.0029915366737395437,
+ "loss": 6.613471031188965,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.02084030583500862,
+ "learning_rate": 0.0029913939734359306,
+ "loss": 6.559704780578613,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.01327502354979515,
+ "learning_rate": 0.00299125008362836,
+ "loss": 6.564910888671875,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.015091734007000923,
+ "learning_rate": 0.0029911050044327614,
+ "loss": 6.571262836456299,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.015463982708752155,
+ "learning_rate": 0.002990958735966023,
+ "loss": 6.565750598907471,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.013636826537549496,
+ "learning_rate": 0.0029908112783459894,
+ "loss": 6.57981014251709,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.010613273829221725,
+ "learning_rate": 0.0029906626316914655,
+ "loss": 6.538505554199219,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.01804237812757492,
+ "learning_rate": 0.0029905127961222127,
+ "loss": 6.5282511711120605,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.015131509862840176,
+ "learning_rate": 0.0029903617717589506,
+ "loss": 6.531335830688477,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.010518206283450127,
+ "learning_rate": 0.002990209558723358,
+ "loss": 6.521886825561523,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.011360022239387035,
+ "learning_rate": 0.002990056157138069,
+ "loss": 6.528301239013672,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.011556181125342846,
+ "learning_rate": 0.0029899015671266773,
+ "loss": 6.497917652130127,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.01365791168063879,
+ "learning_rate": 0.0029897457888137325,
+ "loss": 6.52849006652832,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.01091043371707201,
+ "learning_rate": 0.0029895888223247426,
+ "loss": 6.492197036743164,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.008484847843647003,
+ "learning_rate": 0.002989430667786173,
+ "loss": 6.497377395629883,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.00836879201233387,
+ "learning_rate": 0.0029892713253254466,
+ "loss": 6.495189666748047,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.009781497530639172,
+ "learning_rate": 0.0029891107950709406,
+ "loss": 6.454085826873779,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.009603648446500301,
+ "learning_rate": 0.0029889490771519934,
+ "loss": 6.499475479125977,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.007991855964064598,
+ "learning_rate": 0.0029887861716988975,
+ "loss": 6.460785865783691,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.008447187952697277,
+ "learning_rate": 0.0029886220788429025,
+ "loss": 6.504582405090332,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.010945498012006283,
+ "learning_rate": 0.0029884567987162156,
+ "loss": 6.49589729309082,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.011096215806901455,
+ "learning_rate": 0.0029882903314519997,
+ "loss": 6.460657119750977,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.013094739057123661,
+ "learning_rate": 0.0029881226771843747,
+ "loss": 6.462450981140137,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.015423037111759186,
+ "learning_rate": 0.0029879538360484165,
+ "loss": 6.45565128326416,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.0182222630828619,
+ "learning_rate": 0.002987783808180157,
+ "loss": 6.452227592468262,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.01981227844953537,
+ "learning_rate": 0.002987612593716585,
+ "loss": 6.477701663970947,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.026733700186014175,
+ "learning_rate": 0.002987440192795645,
+ "loss": 6.452212333679199,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.03504155948758125,
+ "learning_rate": 0.0029872666055562372,
+ "loss": 6.509665489196777,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.027970125898718834,
+ "learning_rate": 0.002987091832138217,
+ "loss": 6.487089157104492,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.013714591972529888,
+ "learning_rate": 0.0029869158726823974,
+ "loss": 6.469569206237793,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.016472695395350456,
+ "learning_rate": 0.0029867387273305453,
+ "loss": 6.461688995361328,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.018457358703017235,
+ "learning_rate": 0.0029865603962253833,
+ "loss": 6.490985870361328,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.011045100167393684,
+ "learning_rate": 0.0029863808795105896,
+ "loss": 6.403701305389404,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.014126897789537907,
+ "learning_rate": 0.0029862001773307972,
+ "loss": 6.379804611206055,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.014015335589647293,
+ "learning_rate": 0.0029860182898315954,
+ "loss": 6.404377460479736,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.010032658465206623,
+ "learning_rate": 0.002985835217159527,
+ "loss": 6.40291690826416,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.015469838865101337,
+ "learning_rate": 0.0029856509594620912,
+ "loss": 6.4414520263671875,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.011106910184025764,
+ "learning_rate": 0.0029854655168877397,
+ "loss": 6.396444320678711,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.01262320950627327,
+ "learning_rate": 0.0029852788895858816,
+ "loss": 6.385222434997559,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.012783976271748543,
+ "learning_rate": 0.002985091077706878,
+ "loss": 6.3818135261535645,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.01224067248404026,
+ "learning_rate": 0.0029849020814020468,
+ "loss": 6.37886905670166,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.010540086776018143,
+ "learning_rate": 0.0029847119008236576,
+ "loss": 6.37799072265625,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.012479913420975208,
+ "learning_rate": 0.0029845205361249366,
+ "loss": 6.365094184875488,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.010080077685415745,
+ "learning_rate": 0.0029843279874600622,
+ "loss": 6.371852874755859,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.013266298919916153,
+ "learning_rate": 0.002984134254984168,
+ "loss": 6.380039215087891,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.01014771405607462,
+ "learning_rate": 0.002983939338853339,
+ "loss": 6.370337963104248,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.008164579048752785,
+ "learning_rate": 0.0029837432392246185,
+ "loss": 6.331899642944336,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.01020442508161068,
+ "learning_rate": 0.002983545956255998,
+ "loss": 6.35077428817749,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.009898579679429531,
+ "learning_rate": 0.002983347490106426,
+ "loss": 6.371985912322998,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.009668919257819653,
+ "learning_rate": 0.002983147840935802,
+ "loss": 6.356776237487793,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.012560366652905941,
+ "learning_rate": 0.0029829470089049815,
+ "loss": 6.357911109924316,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.018435249105095863,
+ "learning_rate": 0.0029827449941757698,
+ "loss": 6.366821765899658,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.029046855866909027,
+ "learning_rate": 0.002982541796910927,
+ "loss": 6.367595672607422,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.043010637164115906,
+ "learning_rate": 0.002982337417274165,
+ "loss": 6.413588523864746,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.045470818877220154,
+ "learning_rate": 0.0029821318554301497,
+ "loss": 6.435015678405762,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.02148726023733616,
+ "learning_rate": 0.002981925111544498,
+ "loss": 6.330894470214844,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.026808688417077065,
+ "learning_rate": 0.00298171718578378,
+ "loss": 6.384839057922363,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.018495213240385056,
+ "learning_rate": 0.002981508078315517,
+ "loss": 6.36277961730957,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.025393597781658173,
+ "learning_rate": 0.0029812977893081834,
+ "loss": 6.3702239990234375,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020856289193034172,
+ "learning_rate": 0.002981086318931206,
+ "loss": 6.354547500610352,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.017827816307544708,
+ "learning_rate": 0.002980873667354962,
+ "loss": 6.339264869689941,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.02639714442193508,
+ "learning_rate": 0.002980659834750781,
+ "loss": 6.3254194259643555,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.015238096937537193,
+ "learning_rate": 0.002980444821290944,
+ "loss": 6.365842819213867,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.012863729149103165,
+ "learning_rate": 0.002980228627148684,
+ "loss": 6.356776714324951,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.014739899896085262,
+ "learning_rate": 0.0029800112524981845,
+ "loss": 6.332481384277344,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.013582171872258186,
+ "learning_rate": 0.0029797926975145797,
+ "loss": 6.311968803405762,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.012704637832939625,
+ "learning_rate": 0.0029795729623739555,
+ "loss": 6.321654319763184,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.012550560757517815,
+ "learning_rate": 0.0029793520472533495,
+ "loss": 6.298409461975098,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.011170165613293648,
+ "learning_rate": 0.002979129952330748,
+ "loss": 6.294830799102783,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.326757907867432,
+ "eval_runtime": 51.6566,
+ "eval_samples_per_second": 47.274,
+ "eval_steps_per_second": 1.491,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.010947446338832378,
+ "learning_rate": 0.0029789066777850897,
+ "loss": 6.286581039428711,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.010097995400428772,
+ "learning_rate": 0.002978682223796262,
+ "loss": 6.288301467895508,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.01224475260823965,
+ "learning_rate": 0.002978456590545104,
+ "loss": 6.264986991882324,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.009734537452459335,
+ "learning_rate": 0.0029782297782134045,
+ "loss": 6.304243087768555,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.01136956550180912,
+ "learning_rate": 0.0029780017869839013,
+ "loss": 6.244938373565674,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.00935293547809124,
+ "learning_rate": 0.002977772617040284,
+ "loss": 6.251852035522461,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.011118784546852112,
+ "learning_rate": 0.0029775422685671894,
+ "loss": 6.256633758544922,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.010837923735380173,
+ "learning_rate": 0.0029773107417502055,
+ "loss": 6.263411045074463,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.011920305900275707,
+ "learning_rate": 0.0029770780367758703,
+ "loss": 6.239311218261719,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.014117483049631119,
+ "learning_rate": 0.0029768441538316685,
+ "loss": 6.253314018249512,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.022502148523926735,
+ "learning_rate": 0.002976609093106036,
+ "loss": 6.246156692504883,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.03348720073699951,
+ "learning_rate": 0.002976372854788357,
+ "loss": 6.255914688110352,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.03427121415734291,
+ "learning_rate": 0.002976135439068965,
+ "loss": 6.289946556091309,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.02407355234026909,
+ "learning_rate": 0.002975896846139141,
+ "loss": 6.288534164428711,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.017861567437648773,
+ "learning_rate": 0.0029756570761911147,
+ "loss": 6.258049964904785,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.019980227574706078,
+ "learning_rate": 0.0029754161294180646,
+ "loss": 6.250895023345947,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.01443127728998661,
+ "learning_rate": 0.0029751740060141177,
+ "loss": 6.212200164794922,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.014975055120885372,
+ "learning_rate": 0.0029749307061743475,
+ "loss": 6.230568885803223,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01408296450972557,
+ "learning_rate": 0.0029746862300947773,
+ "loss": 6.243950366973877,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.011649327352643013,
+ "learning_rate": 0.0029744405779723766,
+ "loss": 6.197744846343994,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.013392188586294651,
+ "learning_rate": 0.002974193750005063,
+ "loss": 6.207277297973633,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.010902997106313705,
+ "learning_rate": 0.002973945746391701,
+ "loss": 6.166373252868652,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.012915392406284809,
+ "learning_rate": 0.0029736965673321032,
+ "loss": 6.186316013336182,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.010772830806672573,
+ "learning_rate": 0.0029734462130270273,
+ "loss": 6.172578811645508,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.010274911299347878,
+ "learning_rate": 0.002973194683678181,
+ "loss": 6.206851005554199,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.013210494071245193,
+ "learning_rate": 0.0029729419794882164,
+ "loss": 6.21115779876709,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.013372313231229782,
+ "learning_rate": 0.0029726881006607326,
+ "loss": 6.18442440032959,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.008105908520519733,
+ "learning_rate": 0.0029724330474002743,
+ "loss": 6.204354286193848,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.014167108573019505,
+ "learning_rate": 0.0029721768199123346,
+ "loss": 6.180662155151367,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.008339803665876389,
+ "learning_rate": 0.002971919418403351,
+ "loss": 6.150079727172852,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.015939386561512947,
+ "learning_rate": 0.0029716608430807073,
+ "loss": 6.175203323364258,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.010777109302580357,
+ "learning_rate": 0.0029714010941527327,
+ "loss": 6.166106224060059,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.011274735443294048,
+ "learning_rate": 0.002971140171828702,
+ "loss": 6.163111686706543,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013626808300614357,
+ "learning_rate": 0.002970878076318836,
+ "loss": 6.140850067138672,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.016907762736082077,
+ "learning_rate": 0.0029706148078343,
+ "loss": 6.192012786865234,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.018817441537976265,
+ "learning_rate": 0.002970350366587205,
+ "loss": 6.162238121032715,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.02829768694937229,
+ "learning_rate": 0.0029700847527906068,
+ "loss": 6.159066200256348,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.02034052088856697,
+ "learning_rate": 0.002969817966658505,
+ "loss": 6.164158344268799,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.01816369965672493,
+ "learning_rate": 0.0029695500084058447,
+ "loss": 6.132316589355469,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.013685261830687523,
+ "learning_rate": 0.0029692808782485146,
+ "loss": 6.152181625366211,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.009791149757802486,
+ "learning_rate": 0.0029690105764033485,
+ "loss": 6.138544082641602,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.014827069826424122,
+ "learning_rate": 0.002968739103088124,
+ "loss": 6.135766983032227,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.01391249243170023,
+ "learning_rate": 0.0029684664585215618,
+ "loss": 6.146172523498535,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.012729899026453495,
+ "learning_rate": 0.002968192642923326,
+ "loss": 6.122613430023193,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.014175458811223507,
+ "learning_rate": 0.002967917656514026,
+ "loss": 6.12574577331543,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.010958840139210224,
+ "learning_rate": 0.002967641499515213,
+ "loss": 6.111088752746582,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.015380816534161568,
+ "learning_rate": 0.0029673641721493817,
+ "loss": 6.100615501403809,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.025339096784591675,
+ "learning_rate": 0.00296708567463997,
+ "loss": 6.139659881591797,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.03695995733141899,
+ "learning_rate": 0.002966806007211358,
+ "loss": 6.160996437072754,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.0408046580851078,
+ "learning_rate": 0.0029665251700888686,
+ "loss": 6.173980712890625,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.025022024288773537,
+ "learning_rate": 0.0029662431634987674,
+ "loss": 6.145794868469238,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.027254154905676842,
+ "learning_rate": 0.002965959987668262,
+ "loss": 6.16928243637085,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.06191903352737427,
+ "learning_rate": 0.0029656756428255017,
+ "loss": 6.138525009155273,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0348956398665905,
+ "learning_rate": 0.002965390129199579,
+ "loss": 6.185477256774902,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.020859677344560623,
+ "learning_rate": 0.002965103447020526,
+ "loss": 6.148846626281738,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.022810814902186394,
+ "learning_rate": 0.0029648155965193174,
+ "loss": 6.1701178550720215,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.017981817945837975,
+ "learning_rate": 0.00296452657792787,
+ "loss": 6.1323771476745605,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.014905866235494614,
+ "learning_rate": 0.00296423639147904,
+ "loss": 6.123426914215088,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.01463468000292778,
+ "learning_rate": 0.002963945037406626,
+ "loss": 6.146122932434082,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.021616583690047264,
+ "learning_rate": 0.002963652515945366,
+ "loss": 6.093459606170654,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.01550968922674656,
+ "learning_rate": 0.00296335882733094,
+ "loss": 6.114929676055908,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.014597428031265736,
+ "learning_rate": 0.0029630639717999675,
+ "loss": 6.113945960998535,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.018249280750751495,
+ "learning_rate": 0.002962767949590008,
+ "loss": 6.130507469177246,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.014160148799419403,
+ "learning_rate": 0.0029624707609395614,
+ "loss": 6.086946487426758,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.015563832595944405,
+ "learning_rate": 0.002962172406088067,
+ "loss": 6.083332538604736,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.012240015901625156,
+ "learning_rate": 0.002961872885275904,
+ "loss": 6.0822906494140625,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.0141651201993227,
+ "learning_rate": 0.002961572198744391,
+ "loss": 6.049562454223633,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.010004645213484764,
+ "learning_rate": 0.0029612703467357866,
+ "loss": 6.067063808441162,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.011790835298597813,
+ "learning_rate": 0.0029609673294932856,
+ "loss": 6.062193870544434,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.011088060215115547,
+ "learning_rate": 0.002960663147261025,
+ "loss": 6.070950031280518,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.010387620888650417,
+ "learning_rate": 0.002960357800284078,
+ "loss": 6.032423973083496,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.012074458412826061,
+ "learning_rate": 0.002960051288808458,
+ "loss": 6.066788673400879,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.097111701965332,
+ "eval_runtime": 51.8742,
+ "eval_samples_per_second": 47.075,
+ "eval_steps_per_second": 1.484,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.013429952785372734,
+ "learning_rate": 0.0029597436130811153,
+ "loss": 6.022172451019287,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.01764705777168274,
+ "learning_rate": 0.0029594347733499387,
+ "loss": 6.029745101928711,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.02403099089860916,
+ "learning_rate": 0.002959124769863755,
+ "loss": 6.04707145690918,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.029030559584498405,
+ "learning_rate": 0.002958813602872328,
+ "loss": 6.05479097366333,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.029980644583702087,
+ "learning_rate": 0.0029585012726263595,
+ "loss": 6.068657875061035,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.021424276754260063,
+ "learning_rate": 0.0029581877793774885,
+ "loss": 6.062008857727051,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.018076395615935326,
+ "learning_rate": 0.002957873123378291,
+ "loss": 6.023279190063477,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.017688792198896408,
+ "learning_rate": 0.002957557304882279,
+ "loss": 6.014641284942627,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.02313626930117607,
+ "learning_rate": 0.0029572403241439026,
+ "loss": 6.026420593261719,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.04045062139630318,
+ "learning_rate": 0.0029569221814185467,
+ "loss": 6.007199764251709,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.015792209655046463,
+ "learning_rate": 0.0029566028769625334,
+ "loss": 6.017209053039551,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.01990479789674282,
+ "learning_rate": 0.002956282411033121,
+ "loss": 5.982841491699219,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.01191043108701706,
+ "learning_rate": 0.002955960783888503,
+ "loss": 6.0035929679870605,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.01732831820845604,
+ "learning_rate": 0.002955637995787807,
+ "loss": 6.025006294250488,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.032948266714811325,
+ "learning_rate": 0.0029553140469910995,
+ "loss": 6.007736682891846,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.011192727833986282,
+ "learning_rate": 0.00295498893775938,
+ "loss": 6.0022687911987305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014646840281784534,
+ "learning_rate": 0.0029546626683545824,
+ "loss": 5.976479530334473,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01589796133339405,
+ "learning_rate": 0.002954335239039576,
+ "loss": 5.980949878692627,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.015129734762012959,
+ "learning_rate": 0.0029540066500781657,
+ "loss": 6.0073394775390625,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.011140518821775913,
+ "learning_rate": 0.002953676901735088,
+ "loss": 5.976587295532227,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.0109480582177639,
+ "learning_rate": 0.0029533459942760174,
+ "loss": 6.001427173614502,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.013346790336072445,
+ "learning_rate": 0.0029530139279675583,
+ "loss": 5.96919059753418,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.012474581599235535,
+ "learning_rate": 0.002952680703077251,
+ "loss": 5.970426559448242,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01140566635876894,
+ "learning_rate": 0.002952346319873569,
+ "loss": 5.989171504974365,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.0137029392644763,
+ "learning_rate": 0.0029520107786259187,
+ "loss": 6.010115623474121,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015517761930823326,
+ "learning_rate": 0.002951674079604639,
+ "loss": 6.003649711608887,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.02173629403114319,
+ "learning_rate": 0.0029513362230810035,
+ "loss": 5.967074871063232,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02683914266526699,
+ "learning_rate": 0.002950997209327216,
+ "loss": 5.994434833526611,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.025865858420729637,
+ "learning_rate": 0.002950657038616415,
+ "loss": 6.010523796081543,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02544897235929966,
+ "learning_rate": 0.002950315711222668,
+ "loss": 5.994078159332275,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.028426019474864006,
+ "learning_rate": 0.0029499732274209777,
+ "loss": 6.020925521850586,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.026174750179052353,
+ "learning_rate": 0.0029496295874872768,
+ "loss": 5.985616683959961,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.02802138403058052,
+ "learning_rate": 0.00294928479169843,
+ "loss": 5.973659038543701,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.019348938018083572,
+ "learning_rate": 0.002948938840332233,
+ "loss": 5.990137577056885,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.016294041648507118,
+ "learning_rate": 0.0029485917336674115,
+ "loss": 5.962082862854004,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.016167785972356796,
+ "learning_rate": 0.002948243471983625,
+ "loss": 5.9594573974609375,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.014119232073426247,
+ "learning_rate": 0.00294789405556146,
+ "loss": 5.894467353820801,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.013805347494781017,
+ "learning_rate": 0.002947543484682436,
+ "loss": 5.964014530181885,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.014496497809886932,
+ "learning_rate": 0.002947191759629001,
+ "loss": 5.989920616149902,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.014813800342381,
+ "learning_rate": 0.0029468388806845333,
+ "loss": 5.9669413566589355,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.01153622381389141,
+ "learning_rate": 0.0029464848481333423,
+ "loss": 5.918018341064453,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.010525977239012718,
+ "learning_rate": 0.0029461296622606647,
+ "loss": 5.892477989196777,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.011410463601350784,
+ "learning_rate": 0.002945773323352667,
+ "loss": 5.925249099731445,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.010949424467980862,
+ "learning_rate": 0.0029454158316964464,
+ "loss": 5.90645694732666,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.00962804164737463,
+ "learning_rate": 0.002945057187580026,
+ "loss": 5.917054176330566,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.011517546139657497,
+ "learning_rate": 0.0029446973912923597,
+ "loss": 5.919445514678955,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01567666605114937,
+ "learning_rate": 0.0029443364431233285,
+ "loss": 5.884111404418945,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.016781140118837357,
+ "learning_rate": 0.002943974343363742,
+ "loss": 5.9252119064331055,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.02244546078145504,
+ "learning_rate": 0.0029436110923053375,
+ "loss": 5.933143615722656,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.028186792507767677,
+ "learning_rate": 0.00294324669024078,
+ "loss": 5.9079132080078125,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.028585223481059074,
+ "learning_rate": 0.002942881137463661,
+ "loss": 5.925262451171875,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.02781485952436924,
+ "learning_rate": 0.0029425144342685003,
+ "loss": 5.963315010070801,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.021486928686499596,
+ "learning_rate": 0.002942146580950744,
+ "loss": 5.930115699768066,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.016890818253159523,
+ "learning_rate": 0.002941777577806765,
+ "loss": 5.921439170837402,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.0169126745313406,
+ "learning_rate": 0.0029414074251338618,
+ "loss": 5.928220272064209,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.014659176580607891,
+ "learning_rate": 0.002941036123230261,
+ "loss": 5.934546947479248,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.015337883494794369,
+ "learning_rate": 0.002940663672395112,
+ "loss": 5.890145301818848,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01254988368600607,
+ "learning_rate": 0.002940290072928493,
+ "loss": 5.875492095947266,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.014658362604677677,
+ "learning_rate": 0.0029399153251314066,
+ "loss": 5.901272773742676,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.017342250794172287,
+ "learning_rate": 0.002939539429305779,
+ "loss": 5.875853538513184,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.016825608909130096,
+ "learning_rate": 0.0029391623857544643,
+ "loss": 5.941972732543945,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.016716884449124336,
+ "learning_rate": 0.002938784194781238,
+ "loss": 5.912367343902588,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.021707067266106606,
+ "learning_rate": 0.002938404856690803,
+ "loss": 5.895319938659668,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.030519135296344757,
+ "learning_rate": 0.002938024371788785,
+ "loss": 5.897165298461914,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.02572188340127468,
+ "learning_rate": 0.002937642740381733,
+ "loss": 5.879164695739746,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.016286717727780342,
+ "learning_rate": 0.002937259962777122,
+ "loss": 5.841421127319336,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.020656082779169083,
+ "learning_rate": 0.002936876039283347,
+ "loss": 5.9222259521484375,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.015350459143519402,
+ "learning_rate": 0.0029364909702097298,
+ "loss": 5.885847091674805,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01708570495247841,
+ "learning_rate": 0.0029361047558665136,
+ "loss": 5.872044563293457,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.013537376187741756,
+ "learning_rate": 0.0029357173965648634,
+ "loss": 5.866434097290039,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.013208637945353985,
+ "learning_rate": 0.002935328892616868,
+ "loss": 5.8624491691589355,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.012249554507434368,
+ "learning_rate": 0.002934939244335538,
+ "loss": 5.833232879638672,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.90985107421875,
+ "eval_runtime": 51.0696,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.01382362935692072,
+ "learning_rate": 0.0029345484520348062,
+ "loss": 5.835695266723633,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.012947693467140198,
+ "learning_rate": 0.0029341565160295262,
+ "loss": 5.826190948486328,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.013653872534632683,
+ "learning_rate": 0.002933763436635475,
+ "loss": 5.831789016723633,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.027596980333328247,
+ "learning_rate": 0.0029333692141693484,
+ "loss": 5.839963436126709,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01637990213930607,
+ "learning_rate": 0.0029329738489487648,
+ "loss": 5.805248260498047,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.02169182151556015,
+ "learning_rate": 0.0029325773412922627,
+ "loss": 5.838248252868652,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.03518201410770416,
+ "learning_rate": 0.0029321796915193015,
+ "loss": 5.834336280822754,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04983877018094063,
+ "learning_rate": 0.0029317808999502603,
+ "loss": 5.879856109619141,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05085664615035057,
+ "learning_rate": 0.0029313809669064377,
+ "loss": 5.946531295776367,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.03314858302474022,
+ "learning_rate": 0.0029309798927100532,
+ "loss": 5.909746170043945,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.034118980169296265,
+ "learning_rate": 0.0029305776776842453,
+ "loss": 5.903947830200195,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.03401247411966324,
+ "learning_rate": 0.0029301743221530705,
+ "loss": 5.936162948608398,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.043827857822179794,
+ "learning_rate": 0.0029297698264415047,
+ "loss": 5.953949451446533,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.06281330436468124,
+ "learning_rate": 0.002929364190875444,
+ "loss": 5.988225936889648,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.07216024398803711,
+ "learning_rate": 0.0029289574157817014,
+ "loss": 6.187741279602051,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.055205658078193665,
+ "learning_rate": 0.002928549501488008,
+ "loss": 6.177831649780273,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.04834286496043205,
+ "learning_rate": 0.0029281404483230133,
+ "loss": 6.209604263305664,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.0447777584195137,
+ "learning_rate": 0.0029277302566162837,
+ "loss": 6.146507263183594,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.05729430541396141,
+ "learning_rate": 0.0029273189266983035,
+ "loss": 6.280686378479004,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.0410807766020298,
+ "learning_rate": 0.002926906458900473,
+ "loss": 6.246824264526367,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.044607777148485184,
+ "learning_rate": 0.0029264928535551104,
+ "loss": 6.19956111907959,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.05141043663024902,
+ "learning_rate": 0.0029260781109954515,
+ "loss": 6.118578910827637,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.029248764738440514,
+ "learning_rate": 0.0029256622315556448,
+ "loss": 6.155360221862793,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.028403814882040024,
+ "learning_rate": 0.0029252452155707582,
+ "loss": 6.097352027893066,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.025872666388750076,
+ "learning_rate": 0.002924827063376773,
+ "loss": 6.072485446929932,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.018447555601596832,
+ "learning_rate": 0.0029244077753105874,
+ "loss": 6.077045440673828,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.023166058585047722,
+ "learning_rate": 0.0029239873517100147,
+ "loss": 6.040050983428955,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.019579261541366577,
+ "learning_rate": 0.002923565792913782,
+ "loss": 6.039424896240234,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.018369073048233986,
+ "learning_rate": 0.002923143099261531,
+ "loss": 6.021180152893066,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.014188461937010288,
+ "learning_rate": 0.0029227192710938195,
+ "loss": 5.9668049812316895,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.019567398354411125,
+ "learning_rate": 0.0029222943087521182,
+ "loss": 5.989189147949219,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.014190621674060822,
+ "learning_rate": 0.0029218682125788102,
+ "loss": 5.942726135253906,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01375526562333107,
+ "learning_rate": 0.002921440982917195,
+ "loss": 5.942777633666992,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.014203300699591637,
+ "learning_rate": 0.0029210126201114836,
+ "loss": 5.927928924560547,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.014231069944798946,
+ "learning_rate": 0.0029205831245067994,
+ "loss": 5.923510551452637,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.012771515175700188,
+ "learning_rate": 0.0029201524964491796,
+ "loss": 5.908803939819336,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.013399476185441017,
+ "learning_rate": 0.0029197207362855735,
+ "loss": 5.890079021453857,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.025862423703074455,
+ "learning_rate": 0.0029192878443638427,
+ "loss": 5.919907569885254,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.011315986514091492,
+ "learning_rate": 0.0029188538210327595,
+ "loss": 5.901529788970947,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.01235947571694851,
+ "learning_rate": 0.00291841866664201,
+ "loss": 5.903595924377441,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.012175127863883972,
+ "learning_rate": 0.0029179823815421893,
+ "loss": 5.868306636810303,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01195445004850626,
+ "learning_rate": 0.0029175449660848052,
+ "loss": 5.899898529052734,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.012801481410861015,
+ "learning_rate": 0.0029171064206222745,
+ "loss": 5.890621662139893,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.011909718625247478,
+ "learning_rate": 0.002916666745507926,
+ "loss": 5.86421012878418,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01162486057728529,
+ "learning_rate": 0.0029162259410959977,
+ "loss": 5.869139671325684,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.010923407040536404,
+ "learning_rate": 0.0029157840077416374,
+ "loss": 5.869789123535156,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.011943561024963856,
+ "learning_rate": 0.002915340945800904,
+ "loss": 5.873437881469727,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.010295520536601543,
+ "learning_rate": 0.002914896755630763,
+ "loss": 5.819953918457031,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.0093373479321599,
+ "learning_rate": 0.002914451437589092,
+ "loss": 5.829481601715088,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.01182442344725132,
+ "learning_rate": 0.002914004992034674,
+ "loss": 5.855249404907227,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.00939213763922453,
+ "learning_rate": 0.0029135574193272034,
+ "loss": 5.83876371383667,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.011648556217551231,
+ "learning_rate": 0.0029131087198272813,
+ "loss": 5.824906826019287,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.011227169074118137,
+ "learning_rate": 0.002912658893896416,
+ "loss": 5.847869873046875,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.011084851808845997,
+ "learning_rate": 0.002912207941897025,
+ "loss": 5.839183807373047,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.009923825040459633,
+ "learning_rate": 0.002911755864192432,
+ "loss": 5.818821907043457,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.011300318874418736,
+ "learning_rate": 0.002911302661146867,
+ "loss": 5.819517135620117,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.009915645234286785,
+ "learning_rate": 0.0029108483331254684,
+ "loss": 5.776939392089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.008965205401182175,
+ "learning_rate": 0.00291039288049428,
+ "loss": 5.824148178100586,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.011025927029550076,
+ "learning_rate": 0.0029099363036202515,
+ "loss": 5.832704544067383,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.010143411345779896,
+ "learning_rate": 0.0029094786028712387,
+ "loss": 5.802353382110596,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.010021256282925606,
+ "learning_rate": 0.002909019778616003,
+ "loss": 5.802136421203613,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.011356706731021404,
+ "learning_rate": 0.002908559831224211,
+ "loss": 5.807962417602539,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.010525298304855824,
+ "learning_rate": 0.002908098761066433,
+ "loss": 5.787661552429199,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.011147079057991505,
+ "learning_rate": 0.0029076365685141463,
+ "loss": 5.801015377044678,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.008777810260653496,
+ "learning_rate": 0.00290717325393973,
+ "loss": 5.779858589172363,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.011423609219491482,
+ "learning_rate": 0.002906708817716469,
+ "loss": 5.777166843414307,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.008639856241643429,
+ "learning_rate": 0.00290624326021855,
+ "loss": 5.747189521789551,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.009628898464143276,
+ "learning_rate": 0.002905776581821066,
+ "loss": 5.7666916847229,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.011988377198576927,
+ "learning_rate": 0.0029053087829000097,
+ "loss": 5.777951240539551,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.01693224348127842,
+ "learning_rate": 0.00290483986383228,
+ "loss": 5.818393707275391,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.020264796912670135,
+ "learning_rate": 0.002904369824995674,
+ "loss": 5.747611045837402,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.018283747136592865,
+ "learning_rate": 0.0029038986667688964,
+ "loss": 5.789663314819336,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.825676918029785,
+ "eval_runtime": 57.6908,
+ "eval_samples_per_second": 42.329,
+ "eval_steps_per_second": 1.335,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.011447477154433727,
+ "learning_rate": 0.002903426389531549,
+ "loss": 5.724975109100342,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.012539781630039215,
+ "learning_rate": 0.002902952993664138,
+ "loss": 5.747196674346924,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.012179837562143803,
+ "learning_rate": 0.002902478479548069,
+ "loss": 5.748995780944824,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.019896456971764565,
+ "learning_rate": 0.00290200284756565,
+ "loss": 5.751996994018555,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.022261742502450943,
+ "learning_rate": 0.002901526098100089,
+ "loss": 5.7516984939575195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02683536894619465,
+ "learning_rate": 0.0029010482315354946,
+ "loss": 5.716037750244141,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.018057072535157204,
+ "learning_rate": 0.0029005692482568747,
+ "loss": 5.778608322143555,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.011646891944110394,
+ "learning_rate": 0.002900089148650138,
+ "loss": 5.7308268547058105,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02230495773255825,
+ "learning_rate": 0.002899607933102092,
+ "loss": 5.768552780151367,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.01212464738637209,
+ "learning_rate": 0.002899125602000443,
+ "loss": 5.733015060424805,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.011764907278120518,
+ "learning_rate": 0.0028986421557337973,
+ "loss": 5.748104095458984,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.01593147963285446,
+ "learning_rate": 0.002898157594691658,
+ "loss": 5.701960563659668,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.016333773732185364,
+ "learning_rate": 0.0028976719192644277,
+ "loss": 5.716808795928955,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.017547868192195892,
+ "learning_rate": 0.002897185129843406,
+ "loss": 5.724377632141113,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.01662801019847393,
+ "learning_rate": 0.002896697226820791,
+ "loss": 5.737432956695557,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.015781691297888756,
+ "learning_rate": 0.002896208210589676,
+ "loss": 5.7376203536987305,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.014363214373588562,
+ "learning_rate": 0.0028957180815440534,
+ "loss": 5.7342610359191895,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.015048054046928883,
+ "learning_rate": 0.0028952268400788115,
+ "loss": 5.715073585510254,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.013753226026892662,
+ "learning_rate": 0.002894734486589734,
+ "loss": 5.73264217376709,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.015562563203275204,
+ "learning_rate": 0.002894241021473502,
+ "loss": 5.7072038650512695,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.01520819216966629,
+ "learning_rate": 0.0028937464451276916,
+ "loss": 5.683300018310547,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.013024724088609219,
+ "learning_rate": 0.002893250757950773,
+ "loss": 5.726273536682129,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.012760856188833714,
+ "learning_rate": 0.0028927539603421128,
+ "loss": 5.7249064445495605,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.013542579486966133,
+ "learning_rate": 0.0028922560527019723,
+ "loss": 5.70849084854126,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.014801754616200924,
+ "learning_rate": 0.002891757035431506,
+ "loss": 5.725185394287109,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.013000023551285267,
+ "learning_rate": 0.0028912569089327637,
+ "loss": 5.705799102783203,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.012919292785227299,
+ "learning_rate": 0.002890755673608688,
+ "loss": 5.674324989318848,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.015904201194643974,
+ "learning_rate": 0.002890253329863115,
+ "loss": 5.680788993835449,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.015467721968889236,
+ "learning_rate": 0.0028897498781007744,
+ "loss": 5.707531929016113,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.015157260000705719,
+ "learning_rate": 0.002889245318727288,
+ "loss": 5.689894676208496,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.014204306527972221,
+ "learning_rate": 0.00288873965214917,
+ "loss": 5.713001251220703,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.022045845165848732,
+ "learning_rate": 0.0028882328787738273,
+ "loss": 5.657631874084473,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.013380175456404686,
+ "learning_rate": 0.0028877249990095573,
+ "loss": 5.702192306518555,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.016698408871889114,
+ "learning_rate": 0.0028872160132655503,
+ "loss": 5.695206642150879,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.020001472905278206,
+ "learning_rate": 0.0028867059219518866,
+ "loss": 5.731742858886719,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.027350103482604027,
+ "learning_rate": 0.0028861947254795375,
+ "loss": 5.712314605712891,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.021862946450710297,
+ "learning_rate": 0.002885682424260365,
+ "loss": 5.697043418884277,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.019216977059841156,
+ "learning_rate": 0.0028851690187071204,
+ "loss": 5.698813438415527,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.02131095714867115,
+ "learning_rate": 0.002884654509233446,
+ "loss": 5.6918864250183105,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.016583887860178947,
+ "learning_rate": 0.0028841388962538724,
+ "loss": 5.711456298828125,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.016998684033751488,
+ "learning_rate": 0.0028836221801838194,
+ "loss": 5.675747871398926,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.01705755852162838,
+ "learning_rate": 0.0028831043614395966,
+ "loss": 5.697177410125732,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.014604352414608002,
+ "learning_rate": 0.0028825854404384,
+ "loss": 5.681950092315674,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.013561096973717213,
+ "learning_rate": 0.0028820654175983162,
+ "loss": 5.67357063293457,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.021149173378944397,
+ "learning_rate": 0.0028815442933383174,
+ "loss": 5.709477424621582,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.01431041955947876,
+ "learning_rate": 0.0028810220680782644,
+ "loss": 5.7012200355529785,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.01429709978401661,
+ "learning_rate": 0.0028804987422389045,
+ "loss": 5.689794540405273,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.032025713473558426,
+ "learning_rate": 0.0028799743162418712,
+ "loss": 5.6713762283325195,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.013168714940547943,
+ "learning_rate": 0.002879448790509686,
+ "loss": 5.685478210449219,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.014088666997849941,
+ "learning_rate": 0.0028789221654657556,
+ "loss": 5.645514965057373,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.013320053927600384,
+ "learning_rate": 0.002878394441534372,
+ "loss": 5.678678512573242,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.022180087864398956,
+ "learning_rate": 0.002877865619140712,
+ "loss": 5.660525321960449,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.01271889265626669,
+ "learning_rate": 0.002877335698710839,
+ "loss": 5.652164459228516,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.011617431417107582,
+ "learning_rate": 0.002876804680671701,
+ "loss": 5.6389946937561035,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.012040726840496063,
+ "learning_rate": 0.0028762725654511282,
+ "loss": 5.628744602203369,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.013148069381713867,
+ "learning_rate": 0.0028757393534778377,
+ "loss": 5.69610595703125,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.01826276071369648,
+ "learning_rate": 0.0028752050451814276,
+ "loss": 5.634238243103027,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.020393460988998413,
+ "learning_rate": 0.0028746696409923805,
+ "loss": 5.656968593597412,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.017072148621082306,
+ "learning_rate": 0.0028741331413420627,
+ "loss": 5.668219089508057,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.01694098301231861,
+ "learning_rate": 0.0028735955466627217,
+ "loss": 5.670073509216309,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.022760670632123947,
+ "learning_rate": 0.002873056857387488,
+ "loss": 5.649288177490234,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.03613442927598953,
+ "learning_rate": 0.0028725170739503734,
+ "loss": 5.723211288452148,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.04232329502701759,
+ "learning_rate": 0.002871976196786272,
+ "loss": 5.757920742034912,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.044210851192474365,
+ "learning_rate": 0.002871434226330959,
+ "loss": 5.702426910400391,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.044886402785778046,
+ "learning_rate": 0.0028708911630210905,
+ "loss": 5.742881774902344,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.022805195301771164,
+ "learning_rate": 0.002870347007294202,
+ "loss": 5.716353416442871,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02219395525753498,
+ "learning_rate": 0.00286980175958871,
+ "loss": 5.68503475189209,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.019240057095885277,
+ "learning_rate": 0.002869255420343911,
+ "loss": 5.700656890869141,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.01957295648753643,
+ "learning_rate": 0.0028687079899999807,
+ "loss": 5.70595645904541,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016326555982232094,
+ "learning_rate": 0.0028681594689979736,
+ "loss": 5.697229385375977,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015815885737538338,
+ "learning_rate": 0.0028676098577798234,
+ "loss": 5.7020792961120605,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.015865175053477287,
+ "learning_rate": 0.002867059156788342,
+ "loss": 5.648272514343262,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.7417449951171875,
+ "eval_runtime": 51.2263,
+ "eval_samples_per_second": 47.671,
+ "eval_steps_per_second": 1.503,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.016063183546066284,
+ "learning_rate": 0.0028665073664672185,
+ "loss": 5.65411376953125,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.014323881827294827,
+ "learning_rate": 0.0028659544872610214,
+ "loss": 5.629649639129639,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.014005238190293312,
+ "learning_rate": 0.002865400519615195,
+ "loss": 5.655135631561279,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.01385372132062912,
+ "learning_rate": 0.0028648454639760615,
+ "loss": 5.632213115692139,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.014073756523430347,
+ "learning_rate": 0.0028642893207908187,
+ "loss": 5.626431465148926,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.019551943987607956,
+ "learning_rate": 0.0028637320905075406,
+ "loss": 5.595447063446045,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.017571093514561653,
+ "learning_rate": 0.002863173773575179,
+ "loss": 5.661581039428711,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.01691303588449955,
+ "learning_rate": 0.0028626143704435585,
+ "loss": 5.641129493713379,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.01584242656826973,
+ "learning_rate": 0.0028620538815633807,
+ "loss": 5.640626907348633,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.014592235907912254,
+ "learning_rate": 0.0028614923073862213,
+ "loss": 5.611204147338867,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.016636405140161514,
+ "learning_rate": 0.0028609296483645304,
+ "loss": 5.614716053009033,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.01370833721011877,
+ "learning_rate": 0.002860365904951632,
+ "loss": 5.633011817932129,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.013218311592936516,
+ "learning_rate": 0.002859801077601725,
+ "loss": 5.588958740234375,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.016971616074442863,
+ "learning_rate": 0.0028592351667698786,
+ "loss": 5.635064125061035,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.015524905174970627,
+ "learning_rate": 0.0028586681729120393,
+ "loss": 5.611454963684082,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.014769157394766808,
+ "learning_rate": 0.0028581000964850215,
+ "loss": 5.62175178527832,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.014313343912363052,
+ "learning_rate": 0.002857530937946516,
+ "loss": 5.603628158569336,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.016748322173953056,
+ "learning_rate": 0.002856960697755083,
+ "loss": 5.617765426635742,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.013503300957381725,
+ "learning_rate": 0.0028563893763701538,
+ "loss": 5.636135578155518,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.012275655753910542,
+ "learning_rate": 0.002855816974252033,
+ "loss": 5.552533149719238,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.01288773212581873,
+ "learning_rate": 0.002855243491861894,
+ "loss": 5.620527267456055,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.013687008060514927,
+ "learning_rate": 0.002854668929661781,
+ "loss": 5.618595123291016,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.015678439289331436,
+ "learning_rate": 0.0028540932881146092,
+ "loss": 5.604031562805176,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.013567988760769367,
+ "learning_rate": 0.002853516567684162,
+ "loss": 5.549149513244629,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0157297495752573,
+ "learning_rate": 0.002852938768835092,
+ "loss": 5.618505477905273,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.011867019347846508,
+ "learning_rate": 0.0028523598920329233,
+ "loss": 5.579049110412598,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.01398563478142023,
+ "learning_rate": 0.0028517799377440443,
+ "loss": 5.597678184509277,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.015525273978710175,
+ "learning_rate": 0.0028511989064357148,
+ "loss": 5.601663589477539,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.01588466763496399,
+ "learning_rate": 0.002850616798576061,
+ "loss": 5.5650434494018555,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.014276969246566296,
+ "learning_rate": 0.0028500336146340772,
+ "loss": 5.607125282287598,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.013253689743578434,
+ "learning_rate": 0.002849449355079623,
+ "loss": 5.6299824714660645,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.01176661066710949,
+ "learning_rate": 0.0028488640203834263,
+ "loss": 5.560839653015137,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.011712565086781979,
+ "learning_rate": 0.0028482776110170813,
+ "loss": 5.618312835693359,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.013515647500753403,
+ "learning_rate": 0.002847690127453047,
+ "loss": 5.566764831542969,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.020177524536848068,
+ "learning_rate": 0.0028471015701646485,
+ "loss": 5.554571628570557,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.029197178781032562,
+ "learning_rate": 0.0028465119396260757,
+ "loss": 5.6097259521484375,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.027947140857577324,
+ "learning_rate": 0.0028459212363123824,
+ "loss": 5.65598201751709,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.021089809015393257,
+ "learning_rate": 0.0028453294606994887,
+ "loss": 5.67195987701416,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.019936271011829376,
+ "learning_rate": 0.0028447366132641773,
+ "loss": 5.609482765197754,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.0176444910466671,
+ "learning_rate": 0.002844142694484094,
+ "loss": 5.60432767868042,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.023154428228735924,
+ "learning_rate": 0.0028435477048377495,
+ "loss": 5.616270065307617,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.018391480669379234,
+ "learning_rate": 0.002842951644804515,
+ "loss": 5.600588321685791,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.013716557994484901,
+ "learning_rate": 0.002842354514864627,
+ "loss": 5.595095634460449,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.014372860081493855,
+ "learning_rate": 0.002841756315499181,
+ "loss": 5.584517478942871,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.014930845238268375,
+ "learning_rate": 0.002841157047190135,
+ "loss": 5.587005138397217,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.012631502002477646,
+ "learning_rate": 0.0028405567104203095,
+ "loss": 5.594094276428223,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.014677002094686031,
+ "learning_rate": 0.0028399553056733854,
+ "loss": 5.569582939147949,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.015558961778879166,
+ "learning_rate": 0.0028393528334339025,
+ "loss": 5.596535682678223,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.014322095550596714,
+ "learning_rate": 0.0028387492941872624,
+ "loss": 5.570324420928955,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.012913414277136326,
+ "learning_rate": 0.002838144688419726,
+ "loss": 5.573002338409424,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.01251310296356678,
+ "learning_rate": 0.0028375390166184137,
+ "loss": 5.550327777862549,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.013847017660737038,
+ "learning_rate": 0.0028369322792713026,
+ "loss": 5.54567813873291,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.013789358548820019,
+ "learning_rate": 0.0028363244768672313,
+ "loss": 5.571495056152344,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.01236631814390421,
+ "learning_rate": 0.0028357156098958954,
+ "loss": 5.569730758666992,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.013210169039666653,
+ "learning_rate": 0.002835105678847847,
+ "loss": 5.53984260559082,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.01604137383401394,
+ "learning_rate": 0.0028344946842144975,
+ "loss": 5.530598163604736,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.01708815060555935,
+ "learning_rate": 0.002833882626488114,
+ "loss": 5.529256820678711,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.017222166061401367,
+ "learning_rate": 0.0028332695061618197,
+ "loss": 5.564949035644531,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.01803218014538288,
+ "learning_rate": 0.002832655323729595,
+ "loss": 5.560049057006836,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.01658225618302822,
+ "learning_rate": 0.002832040079686276,
+ "loss": 5.553183555603027,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.016834143549203873,
+ "learning_rate": 0.002831423774527554,
+ "loss": 5.573678970336914,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.01856408454477787,
+ "learning_rate": 0.002830806408749974,
+ "loss": 5.539497375488281,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.026909375563263893,
+ "learning_rate": 0.002830187982850936,
+ "loss": 5.551433563232422,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.016046348959207535,
+ "learning_rate": 0.002829568497328696,
+ "loss": 5.566728591918945,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.018582617864012718,
+ "learning_rate": 0.0028289479526823614,
+ "loss": 5.543226718902588,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.0211537666618824,
+ "learning_rate": 0.0028283263494118943,
+ "loss": 5.560711860656738,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.013894101604819298,
+ "learning_rate": 0.0028277036880181095,
+ "loss": 5.528890132904053,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.017581885680556297,
+ "learning_rate": 0.0028270799690026725,
+ "loss": 5.57473087310791,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.01622696965932846,
+ "learning_rate": 0.002826455192868104,
+ "loss": 5.578961372375488,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.01966337114572525,
+ "learning_rate": 0.0028258293601177743,
+ "loss": 5.513942718505859,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021581366658210754,
+ "learning_rate": 0.002825202471255906,
+ "loss": 5.573669910430908,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.020468758419156075,
+ "learning_rate": 0.002824574526787571,
+ "loss": 5.516103744506836,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.626922130584717,
+ "eval_runtime": 54.2404,
+ "eval_samples_per_second": 45.022,
+ "eval_steps_per_second": 1.42,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.014597049914300442,
+ "learning_rate": 0.0028239455272186936,
+ "loss": 5.506453514099121,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.020277054980397224,
+ "learning_rate": 0.002823315473056048,
+ "loss": 5.535350799560547,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.019103925675153732,
+ "learning_rate": 0.0028226843648072562,
+ "loss": 5.522078037261963,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.07158710062503815,
+ "learning_rate": 0.0028220522029807913,
+ "loss": 5.531142234802246,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01958007924258709,
+ "learning_rate": 0.0028214189880859746,
+ "loss": 5.484055042266846,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.0185184795409441,
+ "learning_rate": 0.002820784720632976,
+ "loss": 5.511264801025391,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01976991444826126,
+ "learning_rate": 0.0028201494011328133,
+ "loss": 5.51821231842041,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.02188604325056076,
+ "learning_rate": 0.002819513030097352,
+ "loss": 5.483095169067383,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.019405223429203033,
+ "learning_rate": 0.002818875608039305,
+ "loss": 5.539701461791992,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.016907615587115288,
+ "learning_rate": 0.0028182371354722306,
+ "loss": 5.518251419067383,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.02160538360476494,
+ "learning_rate": 0.0028175976129105364,
+ "loss": 5.509786605834961,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.023125721141695976,
+ "learning_rate": 0.002816957040869473,
+ "loss": 5.535561561584473,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.022233380004763603,
+ "learning_rate": 0.0028163154198651372,
+ "loss": 5.553831100463867,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.017227552831172943,
+ "learning_rate": 0.002815672750414473,
+ "loss": 5.547080993652344,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.017122473567724228,
+ "learning_rate": 0.0028150290330352665,
+ "loss": 5.493778228759766,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.020641913637518883,
+ "learning_rate": 0.0028143842682461498,
+ "loss": 5.536645889282227,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.017383966594934464,
+ "learning_rate": 0.0028137384565665967,
+ "loss": 5.506233215332031,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.017969606444239616,
+ "learning_rate": 0.002813091598516928,
+ "loss": 5.524741172790527,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.01936613954603672,
+ "learning_rate": 0.0028124436946183047,
+ "loss": 5.460989952087402,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.017492495477199554,
+ "learning_rate": 0.0028117947453927313,
+ "loss": 5.487619876861572,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.018786396831274033,
+ "learning_rate": 0.0028111447513630545,
+ "loss": 5.516829013824463,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.020149117335677147,
+ "learning_rate": 0.0028104937130529624,
+ "loss": 5.460541725158691,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.024668103083968163,
+ "learning_rate": 0.0028098416309869857,
+ "loss": 5.497687339782715,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.015888608992099762,
+ "learning_rate": 0.002809188505690494,
+ "loss": 5.494459629058838,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.019327623769640923,
+ "learning_rate": 0.0028085343376896997,
+ "loss": 5.486351490020752,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.026855193078517914,
+ "learning_rate": 0.0028078791275116532,
+ "loss": 5.509825706481934,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.026340370997786522,
+ "learning_rate": 0.002807222875684246,
+ "loss": 5.527161598205566,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.02200835384428501,
+ "learning_rate": 0.002806565582736208,
+ "loss": 5.507146835327148,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.020420534536242485,
+ "learning_rate": 0.0028059072491971085,
+ "loss": 5.5162506103515625,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02493203990161419,
+ "learning_rate": 0.0028052478755973545,
+ "loss": 5.5252556800842285,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.021293973550200462,
+ "learning_rate": 0.0028045874624681925,
+ "loss": 5.523589134216309,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.016593236476182938,
+ "learning_rate": 0.0028039260103417045,
+ "loss": 5.500586032867432,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.018857723101973534,
+ "learning_rate": 0.002803263519750811,
+ "loss": 5.52933406829834,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.015930436551570892,
+ "learning_rate": 0.0028025999912292677,
+ "loss": 5.48813533782959,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.019427936524152756,
+ "learning_rate": 0.002801935425311669,
+ "loss": 5.46573543548584,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.018123053014278412,
+ "learning_rate": 0.0028012698225334427,
+ "loss": 5.499445915222168,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.02035410702228546,
+ "learning_rate": 0.002800603183430853,
+ "loss": 5.513540267944336,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.025838350877165794,
+ "learning_rate": 0.0027999355085409996,
+ "loss": 5.565637588500977,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.029621612280607224,
+ "learning_rate": 0.002799266798401816,
+ "loss": 5.5619215965271,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.028948379680514336,
+ "learning_rate": 0.0027985970535520685,
+ "loss": 5.551331520080566,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.022135930135846138,
+ "learning_rate": 0.0027979262745313604,
+ "loss": 5.527722358703613,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.028884069994091988,
+ "learning_rate": 0.0027972544618801256,
+ "loss": 5.554599761962891,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.02499174140393734,
+ "learning_rate": 0.00279658161613963,
+ "loss": 5.543351173400879,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.021131042391061783,
+ "learning_rate": 0.0027959077378519755,
+ "loss": 5.496039390563965,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.02957489714026451,
+ "learning_rate": 0.002795232827560092,
+ "loss": 5.552300453186035,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.022832119837403297,
+ "learning_rate": 0.0027945568858077433,
+ "loss": 5.488651275634766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.02074701525270939,
+ "learning_rate": 0.002793879913139523,
+ "loss": 5.491244792938232,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.019919803366065025,
+ "learning_rate": 0.002793201910100856,
+ "loss": 5.483855247497559,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.01858169585466385,
+ "learning_rate": 0.002792522877237998,
+ "loss": 5.50787353515625,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02085857465863228,
+ "learning_rate": 0.002791842815098032,
+ "loss": 5.512872695922852,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.01407450158149004,
+ "learning_rate": 0.002791161724228872,
+ "loss": 5.472590446472168,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.01672944612801075,
+ "learning_rate": 0.002790479605179262,
+ "loss": 5.469183921813965,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.016565700992941856,
+ "learning_rate": 0.0027897964584987713,
+ "loss": 5.493677139282227,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.017017360776662827,
+ "learning_rate": 0.0027891122847377996,
+ "loss": 5.464931488037109,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.019842486828565598,
+ "learning_rate": 0.002788427084447574,
+ "loss": 5.461293697357178,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.016210904344916344,
+ "learning_rate": 0.0027877408581801467,
+ "loss": 5.454716682434082,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.02064039371907711,
+ "learning_rate": 0.002787053606488398,
+ "loss": 5.466576099395752,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.018198711797595024,
+ "learning_rate": 0.0027863653299260356,
+ "loss": 5.447747230529785,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.015937626361846924,
+ "learning_rate": 0.00278567602904759,
+ "loss": 5.5055036544799805,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.01687259040772915,
+ "learning_rate": 0.002784985704408419,
+ "loss": 5.422250747680664,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.02622046135365963,
+ "learning_rate": 0.002784294356564705,
+ "loss": 5.466894626617432,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.018672186881303787,
+ "learning_rate": 0.0027836019860734535,
+ "loss": 5.48300838470459,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.02166241593658924,
+ "learning_rate": 0.0027829085934924958,
+ "loss": 5.477658271789551,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018426023423671722,
+ "learning_rate": 0.0027822141793804856,
+ "loss": 5.487186431884766,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.022229891270399094,
+ "learning_rate": 0.0027815187442968995,
+ "loss": 5.488579273223877,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.026930972933769226,
+ "learning_rate": 0.002780822288802037,
+ "loss": 5.492000102996826,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.02733265422284603,
+ "learning_rate": 0.00278012481345702,
+ "loss": 5.4663801193237305,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.02993253618478775,
+ "learning_rate": 0.0027794263188237922,
+ "loss": 5.543368816375732,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03152153640985489,
+ "learning_rate": 0.002778726805465118,
+ "loss": 5.565644264221191,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.031395044177770615,
+ "learning_rate": 0.002778026273944581,
+ "loss": 5.553359508514404,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.026603225618600845,
+ "learning_rate": 0.002777324724826589,
+ "loss": 5.552347183227539,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.024152139201760292,
+ "learning_rate": 0.002776622158676366,
+ "loss": 5.493257522583008,
+ "step": 720
+ },
+ {
+ "epoch": 10.0,
+ "eval_loss": 5.602636814117432,
+ "eval_runtime": 54.6199,
+ "eval_samples_per_second": 44.709,
+ "eval_steps_per_second": 1.41,
+ "step": 720
+ },
+ {
+ "epoch": 10.013973799126637,
+ "grad_norm": 0.02089761756360531,
+ "learning_rate": 0.0027759185760599576,
+ "loss": 5.461747646331787,
+ "step": 721
+ },
+ {
+ "epoch": 10.027947598253276,
+ "grad_norm": 0.018181778490543365,
+ "learning_rate": 0.0027752139775442278,
+ "loss": 5.492741584777832,
+ "step": 722
+ },
+ {
+ "epoch": 10.041921397379912,
+ "grad_norm": 0.020166238769888878,
+ "learning_rate": 0.0027745083636968575,
+ "loss": 5.476003646850586,
+ "step": 723
+ },
+ {
+ "epoch": 10.055895196506551,
+ "grad_norm": 0.015491694211959839,
+ "learning_rate": 0.0027738017350863488,
+ "loss": 5.425764560699463,
+ "step": 724
+ },
+ {
+ "epoch": 10.069868995633188,
+ "grad_norm": 0.022291697561740875,
+ "learning_rate": 0.0027730940922820182,
+ "loss": 5.45866584777832,
+ "step": 725
+ },
+ {
+ "epoch": 10.083842794759825,
+ "grad_norm": 0.018074683845043182,
+ "learning_rate": 0.002772385435854002,
+ "loss": 5.465143203735352,
+ "step": 726
+ },
+ {
+ "epoch": 10.097816593886463,
+ "grad_norm": 0.016872268170118332,
+ "learning_rate": 0.002771675766373251,
+ "loss": 5.446003437042236,
+ "step": 727
+ },
+ {
+ "epoch": 10.1117903930131,
+ "grad_norm": 0.023103339597582817,
+ "learning_rate": 0.0027709650844115336,
+ "loss": 5.411036491394043,
+ "step": 728
+ },
+ {
+ "epoch": 10.125764192139737,
+ "grad_norm": 0.027957158163189888,
+ "learning_rate": 0.0027702533905414324,
+ "loss": 5.452442169189453,
+ "step": 729
+ },
+ {
+ "epoch": 10.139737991266376,
+ "grad_norm": 0.019935399293899536,
+ "learning_rate": 0.002769540685336348,
+ "loss": 5.412145614624023,
+ "step": 730
+ },
+ {
+ "epoch": 10.153711790393013,
+ "grad_norm": 0.014351371675729752,
+ "learning_rate": 0.0027688269693704926,
+ "loss": 5.434524059295654,
+ "step": 731
+ },
+ {
+ "epoch": 10.167685589519651,
+ "grad_norm": 0.0152254244312644,
+ "learning_rate": 0.0027681122432188948,
+ "loss": 5.428544998168945,
+ "step": 732
+ },
+ {
+ "epoch": 10.181659388646288,
+ "grad_norm": 0.015071330592036247,
+ "learning_rate": 0.0027673965074573965,
+ "loss": 5.459929466247559,
+ "step": 733
+ },
+ {
+ "epoch": 10.195633187772925,
+ "grad_norm": 0.023415958508849144,
+ "learning_rate": 0.0027666797626626537,
+ "loss": 5.447159767150879,
+ "step": 734
+ },
+ {
+ "epoch": 10.209606986899564,
+ "grad_norm": 0.017228802666068077,
+ "learning_rate": 0.0027659620094121337,
+ "loss": 5.442273139953613,
+ "step": 735
+ },
+ {
+ "epoch": 10.2235807860262,
+ "grad_norm": 0.014184504747390747,
+ "learning_rate": 0.002765243248284118,
+ "loss": 5.448305130004883,
+ "step": 736
+ },
+ {
+ "epoch": 10.237554585152838,
+ "grad_norm": 0.024384526535868645,
+ "learning_rate": 0.0027645234798576988,
+ "loss": 5.4076337814331055,
+ "step": 737
+ },
+ {
+ "epoch": 10.251528384279476,
+ "grad_norm": 0.043152231723070145,
+ "learning_rate": 0.0027638027047127813,
+ "loss": 5.45344352722168,
+ "step": 738
+ },
+ {
+ "epoch": 10.265502183406113,
+ "grad_norm": 0.02889758162200451,
+ "learning_rate": 0.002763080923430081,
+ "loss": 5.489194393157959,
+ "step": 739
+ },
+ {
+ "epoch": 10.279475982532752,
+ "grad_norm": 0.02686968818306923,
+ "learning_rate": 0.002762358136591123,
+ "loss": 5.493659496307373,
+ "step": 740
+ },
+ {
+ "epoch": 10.293449781659389,
+ "grad_norm": 0.020467868074774742,
+ "learning_rate": 0.002761634344778245,
+ "loss": 5.410489559173584,
+ "step": 741
+ },
+ {
+ "epoch": 10.307423580786025,
+ "grad_norm": 0.019759224727749825,
+ "learning_rate": 0.002760909548574592,
+ "loss": 5.439385890960693,
+ "step": 742
+ },
+ {
+ "epoch": 10.321397379912664,
+ "grad_norm": 0.0233843382447958,
+ "learning_rate": 0.0027601837485641194,
+ "loss": 5.445036888122559,
+ "step": 743
+ },
+ {
+ "epoch": 10.335371179039301,
+ "grad_norm": 0.031753458082675934,
+ "learning_rate": 0.0027594569453315916,
+ "loss": 5.503895282745361,
+ "step": 744
+ },
+ {
+ "epoch": 10.34934497816594,
+ "grad_norm": 0.02089349366724491,
+ "learning_rate": 0.002758729139462581,
+ "loss": 5.45689582824707,
+ "step": 745
+ },
+ {
+ "epoch": 10.363318777292577,
+ "grad_norm": 0.03339357674121857,
+ "learning_rate": 0.002758000331543466,
+ "loss": 5.522361755371094,
+ "step": 746
+ },
+ {
+ "epoch": 10.377292576419213,
+ "grad_norm": 0.019328996539115906,
+ "learning_rate": 0.002757270522161436,
+ "loss": 5.441373825073242,
+ "step": 747
+ },
+ {
+ "epoch": 10.391266375545852,
+ "grad_norm": 0.019863856956362724,
+ "learning_rate": 0.002756539711904484,
+ "loss": 5.4633283615112305,
+ "step": 748
+ },
+ {
+ "epoch": 10.405240174672489,
+ "grad_norm": 0.02141754701733589,
+ "learning_rate": 0.0027558079013614116,
+ "loss": 5.453215599060059,
+ "step": 749
+ },
+ {
+ "epoch": 10.419213973799126,
+ "grad_norm": 0.016591915860772133,
+ "learning_rate": 0.002755075091121824,
+ "loss": 5.463696479797363,
+ "step": 750
+ },
+ {
+ "epoch": 10.433187772925764,
+ "grad_norm": 0.01955028623342514,
+ "learning_rate": 0.0027543412817761345,
+ "loss": 5.411355018615723,
+ "step": 751
+ },
+ {
+ "epoch": 10.447161572052401,
+ "grad_norm": 0.016791662201285362,
+ "learning_rate": 0.0027536064739155596,
+ "loss": 5.459614276885986,
+ "step": 752
+ },
+ {
+ "epoch": 10.46113537117904,
+ "grad_norm": 0.019661981612443924,
+ "learning_rate": 0.00275287066813212,
+ "loss": 5.440699577331543,
+ "step": 753
+ },
+ {
+ "epoch": 10.475109170305677,
+ "grad_norm": 0.01927965320646763,
+ "learning_rate": 0.002752133865018642,
+ "loss": 5.409865856170654,
+ "step": 754
+ },
+ {
+ "epoch": 10.489082969432314,
+ "grad_norm": 0.016483580693602562,
+ "learning_rate": 0.002751396065168754,
+ "loss": 5.390999794006348,
+ "step": 755
+ },
+ {
+ "epoch": 10.503056768558952,
+ "grad_norm": 0.016264773905277252,
+ "learning_rate": 0.0027506572691768888,
+ "loss": 5.418543815612793,
+ "step": 756
+ },
+ {
+ "epoch": 10.51703056768559,
+ "grad_norm": 0.01958460919559002,
+ "learning_rate": 0.00274991747763828,
+ "loss": 5.403289794921875,
+ "step": 757
+ },
+ {
+ "epoch": 10.531004366812226,
+ "grad_norm": 0.01861695572733879,
+ "learning_rate": 0.002749176691148965,
+ "loss": 5.431911468505859,
+ "step": 758
+ },
+ {
+ "epoch": 10.544978165938865,
+ "grad_norm": 0.029594263061881065,
+ "learning_rate": 0.002748434910305782,
+ "loss": 5.399242401123047,
+ "step": 759
+ },
+ {
+ "epoch": 10.558951965065502,
+ "grad_norm": 0.028574613854289055,
+ "learning_rate": 0.0027476921357063697,
+ "loss": 5.440446853637695,
+ "step": 760
+ },
+ {
+ "epoch": 10.57292576419214,
+ "grad_norm": 0.028612585738301277,
+ "learning_rate": 0.0027469483679491693,
+ "loss": 5.410346031188965,
+ "step": 761
+ },
+ {
+ "epoch": 10.586899563318777,
+ "grad_norm": 0.020704898983240128,
+ "learning_rate": 0.00274620360763342,
+ "loss": 5.44694709777832,
+ "step": 762
+ },
+ {
+ "epoch": 10.600873362445414,
+ "grad_norm": 0.020542364567518234,
+ "learning_rate": 0.002745457855359163,
+ "loss": 5.402653694152832,
+ "step": 763
+ },
+ {
+ "epoch": 10.614847161572053,
+ "grad_norm": 0.015762662515044212,
+ "learning_rate": 0.0027447111117272363,
+ "loss": 5.396594047546387,
+ "step": 764
+ },
+ {
+ "epoch": 10.62882096069869,
+ "grad_norm": 0.022096984088420868,
+ "learning_rate": 0.002743963377339278,
+ "loss": 5.398744106292725,
+ "step": 765
+ },
+ {
+ "epoch": 10.642794759825328,
+ "grad_norm": 0.018791764974594116,
+ "learning_rate": 0.002743214652797724,
+ "loss": 5.397758483886719,
+ "step": 766
+ },
+ {
+ "epoch": 10.656768558951965,
+ "grad_norm": 0.01642526686191559,
+ "learning_rate": 0.0027424649387058093,
+ "loss": 5.411459922790527,
+ "step": 767
+ },
+ {
+ "epoch": 10.670742358078602,
+ "grad_norm": 0.017931703478097916,
+ "learning_rate": 0.0027417142356675637,
+ "loss": 5.390264511108398,
+ "step": 768
+ },
+ {
+ "epoch": 10.68471615720524,
+ "grad_norm": 0.018819602206349373,
+ "learning_rate": 0.0027409625442878153,
+ "loss": 5.407539367675781,
+ "step": 769
+ },
+ {
+ "epoch": 10.698689956331878,
+ "grad_norm": 0.019389411434531212,
+ "learning_rate": 0.0027402098651721886,
+ "loss": 5.386474609375,
+ "step": 770
+ },
+ {
+ "epoch": 10.712663755458514,
+ "grad_norm": 0.013645983301103115,
+ "learning_rate": 0.002739456198927104,
+ "loss": 5.373845100402832,
+ "step": 771
+ },
+ {
+ "epoch": 10.726637554585153,
+ "grad_norm": 0.013846895657479763,
+ "learning_rate": 0.002738701546159776,
+ "loss": 5.387155055999756,
+ "step": 772
+ },
+ {
+ "epoch": 10.74061135371179,
+ "grad_norm": 0.013553125783801079,
+ "learning_rate": 0.002737945907478215,
+ "loss": 5.377801895141602,
+ "step": 773
+ },
+ {
+ "epoch": 10.754585152838429,
+ "grad_norm": 0.012771612964570522,
+ "learning_rate": 0.002737189283491225,
+ "loss": 5.344066143035889,
+ "step": 774
+ },
+ {
+ "epoch": 10.768558951965066,
+ "grad_norm": 0.020465301349759102,
+ "learning_rate": 0.002736431674808405,
+ "loss": 5.382771968841553,
+ "step": 775
+ },
+ {
+ "epoch": 10.782532751091702,
+ "grad_norm": 0.014798106625676155,
+ "learning_rate": 0.0027356730820401465,
+ "loss": 5.370522499084473,
+ "step": 776
+ },
+ {
+ "epoch": 10.796506550218341,
+ "grad_norm": 0.01566634140908718,
+ "learning_rate": 0.0027349135057976332,
+ "loss": 5.351980209350586,
+ "step": 777
+ },
+ {
+ "epoch": 10.810480349344978,
+ "grad_norm": 0.017485102638602257,
+ "learning_rate": 0.002734152946692843,
+ "loss": 5.355443477630615,
+ "step": 778
+ },
+ {
+ "epoch": 10.824454148471617,
+ "grad_norm": 0.016646096482872963,
+ "learning_rate": 0.0027333914053385444,
+ "loss": 5.34829044342041,
+ "step": 779
+ },
+ {
+ "epoch": 10.838427947598253,
+ "grad_norm": 0.017293352633714676,
+ "learning_rate": 0.0027326288823482964,
+ "loss": 5.3519487380981445,
+ "step": 780
+ },
+ {
+ "epoch": 10.85240174672489,
+ "grad_norm": 0.018873002380132675,
+ "learning_rate": 0.002731865378336452,
+ "loss": 5.348442077636719,
+ "step": 781
+ },
+ {
+ "epoch": 10.866375545851529,
+ "grad_norm": 0.015504784882068634,
+ "learning_rate": 0.002731100893918151,
+ "loss": 5.341490268707275,
+ "step": 782
+ },
+ {
+ "epoch": 10.880349344978166,
+ "grad_norm": 0.014270870015025139,
+ "learning_rate": 0.0027303354297093254,
+ "loss": 5.359251022338867,
+ "step": 783
+ },
+ {
+ "epoch": 10.894323144104803,
+ "grad_norm": 0.018820680677890778,
+ "learning_rate": 0.002729568986326695,
+ "loss": 5.390280723571777,
+ "step": 784
+ },
+ {
+ "epoch": 10.908296943231441,
+ "grad_norm": 0.021008187904953957,
+ "learning_rate": 0.0027288015643877708,
+ "loss": 5.338380336761475,
+ "step": 785
+ },
+ {
+ "epoch": 10.922270742358078,
+ "grad_norm": 0.016266655176877975,
+ "learning_rate": 0.0027280331645108498,
+ "loss": 5.347043037414551,
+ "step": 786
+ },
+ {
+ "epoch": 10.936244541484717,
+ "grad_norm": 0.01831045374274254,
+ "learning_rate": 0.0027272637873150185,
+ "loss": 5.326537132263184,
+ "step": 787
+ },
+ {
+ "epoch": 10.950218340611354,
+ "grad_norm": 0.02073794975876808,
+ "learning_rate": 0.002726493433420149,
+ "loss": 5.3607025146484375,
+ "step": 788
+ },
+ {
+ "epoch": 10.96419213973799,
+ "grad_norm": 0.022534754127264023,
+ "learning_rate": 0.0027257221034469033,
+ "loss": 5.353292465209961,
+ "step": 789
+ },
+ {
+ "epoch": 10.97816593886463,
+ "grad_norm": 0.021627362817525864,
+ "learning_rate": 0.002724949798016727,
+ "loss": 5.373323917388916,
+ "step": 790
+ },
+ {
+ "epoch": 10.992139737991266,
+ "grad_norm": 0.018226243555545807,
+ "learning_rate": 0.0027241765177518523,
+ "loss": 5.344578266143799,
+ "step": 791
+ },
+ {
+ "epoch": 11.0,
+ "grad_norm": 0.02032875455915928,
+ "learning_rate": 0.002723402263275298,
+ "loss": 5.329159259796143,
+ "step": 792
+ },
+ {
+ "epoch": 11.0,
+ "eval_loss": 5.44969367980957,
+ "eval_runtime": 51.2294,
+ "eval_samples_per_second": 47.668,
+ "eval_steps_per_second": 1.503,
+ "step": 792
+ },
+ {
+ "epoch": 11.013973799126637,
+ "grad_norm": 0.01613701693713665,
+ "learning_rate": 0.0027226270352108664,
+ "loss": 5.3024678230285645,
+ "step": 793
+ },
+ {
+ "epoch": 11.027947598253276,
+ "grad_norm": 0.01940973661839962,
+ "learning_rate": 0.0027218508341831445,
+ "loss": 5.270181179046631,
+ "step": 794
+ },
+ {
+ "epoch": 11.041921397379912,
+ "grad_norm": 0.015840493142604828,
+ "learning_rate": 0.002721073660817504,
+ "loss": 5.261706352233887,
+ "step": 795
+ },
+ {
+ "epoch": 11.055895196506551,
+ "grad_norm": 0.017390675842761993,
+ "learning_rate": 0.0027202955157401,
+ "loss": 5.262207984924316,
+ "step": 796
+ },
+ {
+ "epoch": 11.069868995633188,
+ "grad_norm": 0.015184012241661549,
+ "learning_rate": 0.0027195163995778686,
+ "loss": 5.297155380249023,
+ "step": 797
+ },
+ {
+ "epoch": 11.083842794759825,
+ "grad_norm": 0.013417073525488377,
+ "learning_rate": 0.00271873631295853,
+ "loss": 5.279838562011719,
+ "step": 798
+ },
+ {
+ "epoch": 11.097816593886463,
+ "grad_norm": 0.013024317100644112,
+ "learning_rate": 0.0027179552565105864,
+ "loss": 5.265446186065674,
+ "step": 799
+ },
+ {
+ "epoch": 11.1117903930131,
+ "grad_norm": 0.01426338218152523,
+ "learning_rate": 0.002717173230863321,
+ "loss": 5.284180164337158,
+ "step": 800
+ },
+ {
+ "epoch": 11.125764192139737,
+ "grad_norm": 0.019604040309786797,
+ "learning_rate": 0.002716390236646797,
+ "loss": 5.286281108856201,
+ "step": 801
+ },
+ {
+ "epoch": 11.139737991266376,
+ "grad_norm": 0.019039476290345192,
+ "learning_rate": 0.0027156062744918587,
+ "loss": 5.290433406829834,
+ "step": 802
+ },
+ {
+ "epoch": 11.153711790393013,
+ "grad_norm": 0.019284334033727646,
+ "learning_rate": 0.0027148213450301313,
+ "loss": 5.262433052062988,
+ "step": 803
+ },
+ {
+ "epoch": 11.167685589519651,
+ "grad_norm": 0.0145154669880867,
+ "learning_rate": 0.0027140354488940175,
+ "loss": 5.24387264251709,
+ "step": 804
+ },
+ {
+ "epoch": 11.181659388646288,
+ "grad_norm": 0.014776417054235935,
+ "learning_rate": 0.0027132485867166992,
+ "loss": 5.282435417175293,
+ "step": 805
+ },
+ {
+ "epoch": 11.195633187772925,
+ "grad_norm": 0.012711729854345322,
+ "learning_rate": 0.0027124607591321377,
+ "loss": 5.246735572814941,
+ "step": 806
+ },
+ {
+ "epoch": 11.209606986899564,
+ "grad_norm": 0.013394835405051708,
+ "learning_rate": 0.0027116719667750716,
+ "loss": 5.270739555358887,
+ "step": 807
+ },
+ {
+ "epoch": 11.2235807860262,
+ "grad_norm": 0.013468850404024124,
+ "learning_rate": 0.002710882210281016,
+ "loss": 5.27161169052124,
+ "step": 808
+ },
+ {
+ "epoch": 11.237554585152838,
+ "grad_norm": 0.030262840911746025,
+ "learning_rate": 0.002710091490286264,
+ "loss": 5.256673812866211,
+ "step": 809
+ },
+ {
+ "epoch": 11.251528384279476,
+ "grad_norm": 0.013560288585722446,
+ "learning_rate": 0.0027092998074278846,
+ "loss": 5.253674507141113,
+ "step": 810
+ },
+ {
+ "epoch": 11.265502183406113,
+ "grad_norm": 0.014592813327908516,
+ "learning_rate": 0.002708507162343721,
+ "loss": 5.25804328918457,
+ "step": 811
+ },
+ {
+ "epoch": 11.279475982532752,
+ "grad_norm": 0.014328169636428356,
+ "learning_rate": 0.0027077135556723953,
+ "loss": 5.235117435455322,
+ "step": 812
+ },
+ {
+ "epoch": 11.293449781659389,
+ "grad_norm": 0.01277366653084755,
+ "learning_rate": 0.0027069189880533,
+ "loss": 5.2053985595703125,
+ "step": 813
+ },
+ {
+ "epoch": 11.307423580786025,
+ "grad_norm": 0.012948785908520222,
+ "learning_rate": 0.0027061234601266043,
+ "loss": 5.214653015136719,
+ "step": 814
+ },
+ {
+ "epoch": 11.321397379912664,
+ "grad_norm": 0.014892828650772572,
+ "learning_rate": 0.002705326972533252,
+ "loss": 5.227547645568848,
+ "step": 815
+ },
+ {
+ "epoch": 11.335371179039301,
+ "grad_norm": 0.01427230704575777,
+ "learning_rate": 0.002704529525914957,
+ "loss": 5.241065502166748,
+ "step": 816
+ },
+ {
+ "epoch": 11.34934497816594,
+ "grad_norm": 0.01650116965174675,
+ "learning_rate": 0.0027037311209142092,
+ "loss": 5.227713108062744,
+ "step": 817
+ },
+ {
+ "epoch": 11.363318777292577,
+ "grad_norm": 0.019728578627109528,
+ "learning_rate": 0.0027029317581742677,
+ "loss": 5.2624921798706055,
+ "step": 818
+ },
+ {
+ "epoch": 11.377292576419213,
+ "grad_norm": 0.023847149685025215,
+ "learning_rate": 0.0027021314383391657,
+ "loss": 5.278811454772949,
+ "step": 819
+ },
+ {
+ "epoch": 11.391266375545852,
+ "grad_norm": 0.020690469071269035,
+ "learning_rate": 0.0027013301620537056,
+ "loss": 5.238919258117676,
+ "step": 820
+ },
+ {
+ "epoch": 11.405240174672489,
+ "grad_norm": 0.021211959421634674,
+ "learning_rate": 0.0027005279299634616,
+ "loss": 5.253684997558594,
+ "step": 821
+ },
+ {
+ "epoch": 11.419213973799126,
+ "grad_norm": 0.025854922831058502,
+ "learning_rate": 0.002699724742714778,
+ "loss": 5.2538652420043945,
+ "step": 822
+ },
+ {
+ "epoch": 11.433187772925764,
+ "grad_norm": 0.0215485580265522,
+ "learning_rate": 0.0026989206009547678,
+ "loss": 5.224540710449219,
+ "step": 823
+ },
+ {
+ "epoch": 11.447161572052401,
+ "grad_norm": 0.019827792420983315,
+ "learning_rate": 0.0026981155053313137,
+ "loss": 5.220674514770508,
+ "step": 824
+ },
+ {
+ "epoch": 11.46113537117904,
+ "grad_norm": 0.017689093947410583,
+ "learning_rate": 0.002697309456493066,
+ "loss": 5.228511810302734,
+ "step": 825
+ },
+ {
+ "epoch": 11.475109170305677,
+ "grad_norm": 0.022564806044101715,
+ "learning_rate": 0.002696502455089445,
+ "loss": 5.295933723449707,
+ "step": 826
+ },
+ {
+ "epoch": 11.489082969432314,
+ "grad_norm": 0.020070280879735947,
+ "learning_rate": 0.0026956945017706357,
+ "loss": 5.230920791625977,
+ "step": 827
+ },
+ {
+ "epoch": 11.503056768558952,
+ "grad_norm": 0.025744661688804626,
+ "learning_rate": 0.002694885597187593,
+ "loss": 5.296453952789307,
+ "step": 828
+ },
+ {
+ "epoch": 11.51703056768559,
+ "grad_norm": 0.021263007074594498,
+ "learning_rate": 0.0026940757419920357,
+ "loss": 5.220481872558594,
+ "step": 829
+ },
+ {
+ "epoch": 11.531004366812226,
+ "grad_norm": 0.017577245831489563,
+ "learning_rate": 0.0026932649368364495,
+ "loss": 5.231488227844238,
+ "step": 830
+ },
+ {
+ "epoch": 11.544978165938865,
+ "grad_norm": 0.015562832355499268,
+ "learning_rate": 0.002692453182374086,
+ "loss": 5.175224304199219,
+ "step": 831
+ },
+ {
+ "epoch": 11.558951965065502,
+ "grad_norm": 0.014115403406322002,
+ "learning_rate": 0.0026916404792589606,
+ "loss": 5.228387832641602,
+ "step": 832
+ },
+ {
+ "epoch": 11.57292576419214,
+ "grad_norm": 0.012863830663263798,
+ "learning_rate": 0.002690826828145854,
+ "loss": 5.186727523803711,
+ "step": 833
+ },
+ {
+ "epoch": 11.586899563318777,
+ "grad_norm": 0.014151126146316528,
+ "learning_rate": 0.002690012229690309,
+ "loss": 5.1694817543029785,
+ "step": 834
+ },
+ {
+ "epoch": 11.600873362445414,
+ "grad_norm": 0.01316932775080204,
+ "learning_rate": 0.0026891966845486353,
+ "loss": 5.19931697845459,
+ "step": 835
+ },
+ {
+ "epoch": 11.614847161572053,
+ "grad_norm": 0.011340726166963577,
+ "learning_rate": 0.0026883801933779003,
+ "loss": 5.238619804382324,
+ "step": 836
+ },
+ {
+ "epoch": 11.62882096069869,
+ "grad_norm": 0.012616089545190334,
+ "learning_rate": 0.0026875627568359378,
+ "loss": 5.183376789093018,
+ "step": 837
+ },
+ {
+ "epoch": 11.642794759825328,
+ "grad_norm": 0.012274038046598434,
+ "learning_rate": 0.002686744375581341,
+ "loss": 5.199946403503418,
+ "step": 838
+ },
+ {
+ "epoch": 11.656768558951965,
+ "grad_norm": 0.014990398660302162,
+ "learning_rate": 0.002685925050273465,
+ "loss": 5.168451309204102,
+ "step": 839
+ },
+ {
+ "epoch": 11.670742358078602,
+ "grad_norm": 0.011399757117033005,
+ "learning_rate": 0.0026851047815724255,
+ "loss": 5.173828125,
+ "step": 840
+ },
+ {
+ "epoch": 11.68471615720524,
+ "grad_norm": 0.013638390228152275,
+ "learning_rate": 0.0026842835701390982,
+ "loss": 5.142702102661133,
+ "step": 841
+ },
+ {
+ "epoch": 11.698689956331878,
+ "grad_norm": 0.014472718350589275,
+ "learning_rate": 0.0026834614166351177,
+ "loss": 5.173436164855957,
+ "step": 842
+ },
+ {
+ "epoch": 11.712663755458514,
+ "grad_norm": 0.014509808272123337,
+ "learning_rate": 0.002682638321722879,
+ "loss": 5.173975467681885,
+ "step": 843
+ },
+ {
+ "epoch": 11.726637554585153,
+ "grad_norm": 0.013576583005487919,
+ "learning_rate": 0.0026818142860655352,
+ "loss": 5.1435089111328125,
+ "step": 844
+ },
+ {
+ "epoch": 11.74061135371179,
+ "grad_norm": 0.012914472259581089,
+ "learning_rate": 0.0026809893103269954,
+ "loss": 5.151381492614746,
+ "step": 845
+ },
+ {
+ "epoch": 11.754585152838429,
+ "grad_norm": 0.013803352601826191,
+ "learning_rate": 0.0026801633951719288,
+ "loss": 5.133159637451172,
+ "step": 846
+ },
+ {
+ "epoch": 11.768558951965066,
+ "grad_norm": 0.018488213419914246,
+ "learning_rate": 0.0026793365412657606,
+ "loss": 5.183903694152832,
+ "step": 847
+ },
+ {
+ "epoch": 11.782532751091702,
+ "grad_norm": 0.01843969337642193,
+ "learning_rate": 0.002678508749274671,
+ "loss": 5.1660356521606445,
+ "step": 848
+ },
+ {
+ "epoch": 11.796506550218341,
+ "grad_norm": 0.017181839793920517,
+ "learning_rate": 0.0026776800198655983,
+ "loss": 5.179847717285156,
+ "step": 849
+ },
+ {
+ "epoch": 11.810480349344978,
+ "grad_norm": 0.0148160420358181,
+ "learning_rate": 0.0026768503537062333,
+ "loss": 5.149198532104492,
+ "step": 850
+ },
+ {
+ "epoch": 11.824454148471617,
+ "grad_norm": 0.013251296244561672,
+ "learning_rate": 0.0026760197514650247,
+ "loss": 5.192000389099121,
+ "step": 851
+ },
+ {
+ "epoch": 11.838427947598253,
+ "grad_norm": 0.010746927931904793,
+ "learning_rate": 0.0026751882138111717,
+ "loss": 5.138680458068848,
+ "step": 852
+ },
+ {
+ "epoch": 11.85240174672489,
+ "grad_norm": 0.012316994369029999,
+ "learning_rate": 0.002674355741414631,
+ "loss": 5.169965744018555,
+ "step": 853
+ },
+ {
+ "epoch": 11.866375545851529,
+ "grad_norm": 0.01191797386854887,
+ "learning_rate": 0.0026735223349461106,
+ "loss": 5.128267288208008,
+ "step": 854
+ },
+ {
+ "epoch": 11.880349344978166,
+ "grad_norm": 0.013525084592401981,
+ "learning_rate": 0.0026726879950770698,
+ "loss": 5.124795436859131,
+ "step": 855
+ },
+ {
+ "epoch": 11.894323144104803,
+ "grad_norm": 0.014843370765447617,
+ "learning_rate": 0.002671852722479721,
+ "loss": 5.1497087478637695,
+ "step": 856
+ },
+ {
+ "epoch": 11.908296943231441,
+ "grad_norm": 0.015285412780940533,
+ "learning_rate": 0.00267101651782703,
+ "loss": 5.093118190765381,
+ "step": 857
+ },
+ {
+ "epoch": 11.922270742358078,
+ "grad_norm": 0.013555832207202911,
+ "learning_rate": 0.0026701793817927107,
+ "loss": 5.13249397277832,
+ "step": 858
+ },
+ {
+ "epoch": 11.936244541484717,
+ "grad_norm": 0.010569742880761623,
+ "learning_rate": 0.0026693413150512273,
+ "loss": 5.084077835083008,
+ "step": 859
+ },
+ {
+ "epoch": 11.950218340611354,
+ "grad_norm": 0.011483813636004925,
+ "learning_rate": 0.0026685023182777975,
+ "loss": 5.128049850463867,
+ "step": 860
+ },
+ {
+ "epoch": 11.96419213973799,
+ "grad_norm": 0.013724366202950478,
+ "learning_rate": 0.0026676623921483844,
+ "loss": 5.157539367675781,
+ "step": 861
+ },
+ {
+ "epoch": 11.97816593886463,
+ "grad_norm": 0.014337432570755482,
+ "learning_rate": 0.0026668215373397016,
+ "loss": 5.149948596954346,
+ "step": 862
+ },
+ {
+ "epoch": 11.992139737991266,
+ "grad_norm": 0.014101454056799412,
+ "learning_rate": 0.00266597975452921,
+ "loss": 5.117250442504883,
+ "step": 863
+ },
+ {
+ "epoch": 12.0,
+ "grad_norm": 0.016004368662834167,
+ "learning_rate": 0.0026651370443951197,
+ "loss": 5.143221855163574,
+ "step": 864
+ },
+ {
+ "epoch": 12.0,
+ "eval_loss": 5.249833583831787,
+ "eval_runtime": 51.5543,
+ "eval_samples_per_second": 47.368,
+ "eval_steps_per_second": 1.494,
+ "step": 864
+ },
+ {
+ "epoch": 12.013973799126637,
+ "grad_norm": 0.01297403872013092,
+ "learning_rate": 0.0026642934076163867,
+ "loss": 5.089287281036377,
+ "step": 865
+ },
+ {
+ "epoch": 12.027947598253276,
+ "grad_norm": 0.012867718935012817,
+ "learning_rate": 0.0026634488448727144,
+ "loss": 5.0784101486206055,
+ "step": 866
+ },
+ {
+ "epoch": 12.041921397379912,
+ "grad_norm": 0.014600927010178566,
+ "learning_rate": 0.0026626033568445504,
+ "loss": 5.050220489501953,
+ "step": 867
+ },
+ {
+ "epoch": 12.055895196506551,
+ "grad_norm": 0.016619572415947914,
+ "learning_rate": 0.0026617569442130905,
+ "loss": 5.079303741455078,
+ "step": 868
+ },
+ {
+ "epoch": 12.069868995633188,
+ "grad_norm": 0.015560587868094444,
+ "learning_rate": 0.0026609096076602736,
+ "loss": 5.072385311126709,
+ "step": 869
+ },
+ {
+ "epoch": 12.083842794759825,
+ "grad_norm": 0.01774057187139988,
+ "learning_rate": 0.0026600613478687833,
+ "loss": 5.122529983520508,
+ "step": 870
+ },
+ {
+ "epoch": 12.097816593886463,
+ "grad_norm": 0.019971400499343872,
+ "learning_rate": 0.0026592121655220472,
+ "loss": 5.085966110229492,
+ "step": 871
+ },
+ {
+ "epoch": 12.1117903930131,
+ "grad_norm": 0.017667256295681,
+ "learning_rate": 0.0026583620613042365,
+ "loss": 5.078432559967041,
+ "step": 872
+ },
+ {
+ "epoch": 12.125764192139737,
+ "grad_norm": 0.022369660437107086,
+ "learning_rate": 0.002657511035900264,
+ "loss": 5.0630903244018555,
+ "step": 873
+ },
+ {
+ "epoch": 12.139737991266376,
+ "grad_norm": 0.017964458093047142,
+ "learning_rate": 0.002656659089995787,
+ "loss": 5.0846076011657715,
+ "step": 874
+ },
+ {
+ "epoch": 12.153711790393013,
+ "grad_norm": 0.021769145503640175,
+ "learning_rate": 0.002655806224277201,
+ "loss": 5.081344127655029,
+ "step": 875
+ },
+ {
+ "epoch": 12.167685589519651,
+ "grad_norm": 0.02487620897591114,
+ "learning_rate": 0.0026549524394316464,
+ "loss": 5.106287956237793,
+ "step": 876
+ },
+ {
+ "epoch": 12.181659388646288,
+ "grad_norm": 0.022516218945384026,
+ "learning_rate": 0.0026540977361470012,
+ "loss": 5.128098487854004,
+ "step": 877
+ },
+ {
+ "epoch": 12.195633187772925,
+ "grad_norm": 0.02099926583468914,
+ "learning_rate": 0.0026532421151118845,
+ "loss": 5.103562355041504,
+ "step": 878
+ },
+ {
+ "epoch": 12.209606986899564,
+ "grad_norm": 0.01899387687444687,
+ "learning_rate": 0.002652385577015655,
+ "loss": 5.062675476074219,
+ "step": 879
+ },
+ {
+ "epoch": 12.2235807860262,
+ "grad_norm": 0.016801996156573296,
+ "learning_rate": 0.002651528122548409,
+ "loss": 5.096002578735352,
+ "step": 880
+ },
+ {
+ "epoch": 12.237554585152838,
+ "grad_norm": 0.014009200967848301,
+ "learning_rate": 0.0026506697524009836,
+ "loss": 5.075201034545898,
+ "step": 881
+ },
+ {
+ "epoch": 12.251528384279476,
+ "grad_norm": 0.014653858728706837,
+ "learning_rate": 0.002649810467264951,
+ "loss": 5.090478897094727,
+ "step": 882
+ },
+ {
+ "epoch": 12.265502183406113,
+ "grad_norm": 0.014918150380253792,
+ "learning_rate": 0.0026489502678326226,
+ "loss": 5.070797920227051,
+ "step": 883
+ },
+ {
+ "epoch": 12.279475982532752,
+ "grad_norm": 0.01327288057655096,
+ "learning_rate": 0.0026480891547970443,
+ "loss": 5.030500411987305,
+ "step": 884
+ },
+ {
+ "epoch": 12.293449781659389,
+ "grad_norm": 0.012885140255093575,
+ "learning_rate": 0.002647227128852,
+ "loss": 5.0845537185668945,
+ "step": 885
+ },
+ {
+ "epoch": 12.307423580786025,
+ "grad_norm": 0.012848512269556522,
+ "learning_rate": 0.002646364190692009,
+ "loss": 5.079484939575195,
+ "step": 886
+ },
+ {
+ "epoch": 12.321397379912664,
+ "grad_norm": 0.01149789709597826,
+ "learning_rate": 0.002645500341012324,
+ "loss": 5.036316394805908,
+ "step": 887
+ },
+ {
+ "epoch": 12.335371179039301,
+ "grad_norm": 0.010360757820308208,
+ "learning_rate": 0.0026446355805089337,
+ "loss": 5.04129695892334,
+ "step": 888
+ },
+ {
+ "epoch": 12.34934497816594,
+ "grad_norm": 0.010426290333271027,
+ "learning_rate": 0.00264376990987856,
+ "loss": 5.051270484924316,
+ "step": 889
+ },
+ {
+ "epoch": 12.363318777292577,
+ "grad_norm": 0.01018153689801693,
+ "learning_rate": 0.002642903329818658,
+ "loss": 5.102342128753662,
+ "step": 890
+ },
+ {
+ "epoch": 12.377292576419213,
+ "grad_norm": 0.011879283003509045,
+ "learning_rate": 0.0026420358410274153,
+ "loss": 5.053501129150391,
+ "step": 891
+ },
+ {
+ "epoch": 12.391266375545852,
+ "grad_norm": 0.010418329387903214,
+ "learning_rate": 0.002641167444203752,
+ "loss": 5.030653476715088,
+ "step": 892
+ },
+ {
+ "epoch": 12.405240174672489,
+ "grad_norm": 0.012302013114094734,
+ "learning_rate": 0.0026402981400473202,
+ "loss": 5.0402069091796875,
+ "step": 893
+ },
+ {
+ "epoch": 12.419213973799126,
+ "grad_norm": 0.011392916552722454,
+ "learning_rate": 0.002639427929258502,
+ "loss": 5.024111747741699,
+ "step": 894
+ },
+ {
+ "epoch": 12.433187772925764,
+ "grad_norm": 0.013075058348476887,
+ "learning_rate": 0.0026385568125384114,
+ "loss": 5.021731376647949,
+ "step": 895
+ },
+ {
+ "epoch": 12.447161572052401,
+ "grad_norm": 0.015349832363426685,
+ "learning_rate": 0.0026376847905888903,
+ "loss": 5.03192138671875,
+ "step": 896
+ },
+ {
+ "epoch": 12.46113537117904,
+ "grad_norm": 0.01588309183716774,
+ "learning_rate": 0.0026368118641125113,
+ "loss": 4.966363906860352,
+ "step": 897
+ },
+ {
+ "epoch": 12.475109170305677,
+ "grad_norm": 0.014378590509295464,
+ "learning_rate": 0.0026359380338125764,
+ "loss": 5.070773124694824,
+ "step": 898
+ },
+ {
+ "epoch": 12.489082969432314,
+ "grad_norm": 0.012306655757129192,
+ "learning_rate": 0.002635063300393114,
+ "loss": 5.025704860687256,
+ "step": 899
+ },
+ {
+ "epoch": 12.503056768558952,
+ "grad_norm": 0.01212004292756319,
+ "learning_rate": 0.002634187664558881,
+ "loss": 4.965104103088379,
+ "step": 900
+ },
+ {
+ "epoch": 12.51703056768559,
+ "grad_norm": 0.013538849540054798,
+ "learning_rate": 0.0026333111270153624,
+ "loss": 5.029810905456543,
+ "step": 901
+ },
+ {
+ "epoch": 12.531004366812226,
+ "grad_norm": 0.012052906677126884,
+ "learning_rate": 0.0026324336884687672,
+ "loss": 5.0262908935546875,
+ "step": 902
+ },
+ {
+ "epoch": 12.544978165938865,
+ "grad_norm": 0.011368660256266594,
+ "learning_rate": 0.002631555349626033,
+ "loss": 5.001485824584961,
+ "step": 903
+ },
+ {
+ "epoch": 12.558951965065502,
+ "grad_norm": 0.010215395130217075,
+ "learning_rate": 0.0026306761111948205,
+ "loss": 5.019739151000977,
+ "step": 904
+ },
+ {
+ "epoch": 12.57292576419214,
+ "grad_norm": 0.011449878104031086,
+ "learning_rate": 0.0026297959738835167,
+ "loss": 4.9816083908081055,
+ "step": 905
+ },
+ {
+ "epoch": 12.586899563318777,
+ "grad_norm": 0.012381811626255512,
+ "learning_rate": 0.0026289149384012326,
+ "loss": 5.025032043457031,
+ "step": 906
+ },
+ {
+ "epoch": 12.600873362445414,
+ "grad_norm": 0.00974641740322113,
+ "learning_rate": 0.002628033005457802,
+ "loss": 5.000177383422852,
+ "step": 907
+ },
+ {
+ "epoch": 12.614847161572053,
+ "grad_norm": 0.009515874087810516,
+ "learning_rate": 0.0026271501757637833,
+ "loss": 4.996979713439941,
+ "step": 908
+ },
+ {
+ "epoch": 12.62882096069869,
+ "grad_norm": 0.011125455610454082,
+ "learning_rate": 0.0026262664500304555,
+ "loss": 4.996796607971191,
+ "step": 909
+ },
+ {
+ "epoch": 12.642794759825328,
+ "grad_norm": 0.013290468603372574,
+ "learning_rate": 0.0026253818289698207,
+ "loss": 5.053703308105469,
+ "step": 910
+ },
+ {
+ "epoch": 12.656768558951965,
+ "grad_norm": 0.014650283381342888,
+ "learning_rate": 0.002624496313294602,
+ "loss": 5.001321792602539,
+ "step": 911
+ },
+ {
+ "epoch": 12.670742358078602,
+ "grad_norm": 0.014537827111780643,
+ "learning_rate": 0.002623609903718244,
+ "loss": 5.024584770202637,
+ "step": 912
+ },
+ {
+ "epoch": 12.68471615720524,
+ "grad_norm": 0.012908566743135452,
+ "learning_rate": 0.0026227226009549105,
+ "loss": 5.042878150939941,
+ "step": 913
+ },
+ {
+ "epoch": 12.698689956331878,
+ "grad_norm": 0.01446808222681284,
+ "learning_rate": 0.002621834405719485,
+ "loss": 5.020857810974121,
+ "step": 914
+ },
+ {
+ "epoch": 12.712663755458514,
+ "grad_norm": 0.01535153016448021,
+ "learning_rate": 0.0026209453187275708,
+ "loss": 5.0056304931640625,
+ "step": 915
+ },
+ {
+ "epoch": 12.726637554585153,
+ "grad_norm": 0.01815616339445114,
+ "learning_rate": 0.0026200553406954892,
+ "loss": 5.0467376708984375,
+ "step": 916
+ },
+ {
+ "epoch": 12.74061135371179,
+ "grad_norm": 0.023135516792535782,
+ "learning_rate": 0.0026191644723402795,
+ "loss": 5.033995628356934,
+ "step": 917
+ },
+ {
+ "epoch": 12.754585152838429,
+ "grad_norm": 0.019682930782437325,
+ "learning_rate": 0.002618272714379698,
+ "loss": 5.008149147033691,
+ "step": 918
+ },
+ {
+ "epoch": 12.768558951965066,
+ "grad_norm": 0.018521904945373535,
+ "learning_rate": 0.002617380067532217,
+ "loss": 5.007076263427734,
+ "step": 919
+ },
+ {
+ "epoch": 12.782532751091702,
+ "grad_norm": 0.02118663489818573,
+ "learning_rate": 0.0026164865325170278,
+ "loss": 5.05519962310791,
+ "step": 920
+ },
+ {
+ "epoch": 12.796506550218341,
+ "grad_norm": 0.024684788659214973,
+ "learning_rate": 0.0026155921100540344,
+ "loss": 5.027571201324463,
+ "step": 921
+ },
+ {
+ "epoch": 12.810480349344978,
+ "grad_norm": 0.02609819918870926,
+ "learning_rate": 0.0026146968008638567,
+ "loss": 5.10361385345459,
+ "step": 922
+ },
+ {
+ "epoch": 12.824454148471617,
+ "grad_norm": 0.025958053767681122,
+ "learning_rate": 0.00261380060566783,
+ "loss": 5.0911455154418945,
+ "step": 923
+ },
+ {
+ "epoch": 12.838427947598253,
+ "grad_norm": 0.017067518085241318,
+ "learning_rate": 0.002612903525188002,
+ "loss": 5.045783042907715,
+ "step": 924
+ },
+ {
+ "epoch": 12.85240174672489,
+ "grad_norm": 0.016510486602783203,
+ "learning_rate": 0.0026120055601471345,
+ "loss": 5.069464683532715,
+ "step": 925
+ },
+ {
+ "epoch": 12.866375545851529,
+ "grad_norm": 0.015552333556115627,
+ "learning_rate": 0.0026111067112687012,
+ "loss": 5.024248123168945,
+ "step": 926
+ },
+ {
+ "epoch": 12.880349344978166,
+ "grad_norm": 0.013432685285806656,
+ "learning_rate": 0.0026102069792768886,
+ "loss": 5.004837989807129,
+ "step": 927
+ },
+ {
+ "epoch": 12.894323144104803,
+ "grad_norm": 0.012300201691687107,
+ "learning_rate": 0.0026093063648965943,
+ "loss": 5.031093597412109,
+ "step": 928
+ },
+ {
+ "epoch": 12.908296943231441,
+ "grad_norm": 0.01197744719684124,
+ "learning_rate": 0.002608404868853428,
+ "loss": 4.965237617492676,
+ "step": 929
+ },
+ {
+ "epoch": 12.922270742358078,
+ "grad_norm": 0.01100246049463749,
+ "learning_rate": 0.002607502491873708,
+ "loss": 5.04719352722168,
+ "step": 930
+ },
+ {
+ "epoch": 12.936244541484717,
+ "grad_norm": 0.010111447423696518,
+ "learning_rate": 0.0026065992346844626,
+ "loss": 4.986410617828369,
+ "step": 931
+ },
+ {
+ "epoch": 12.950218340611354,
+ "grad_norm": 0.010669790208339691,
+ "learning_rate": 0.002605695098013431,
+ "loss": 4.9825663566589355,
+ "step": 932
+ },
+ {
+ "epoch": 12.96419213973799,
+ "grad_norm": 0.010805477388203144,
+ "learning_rate": 0.0026047900825890593,
+ "loss": 4.991506576538086,
+ "step": 933
+ },
+ {
+ "epoch": 12.97816593886463,
+ "grad_norm": 0.010886074975132942,
+ "learning_rate": 0.002603884189140502,
+ "loss": 4.987531661987305,
+ "step": 934
+ },
+ {
+ "epoch": 12.992139737991266,
+ "grad_norm": 0.010735839605331421,
+ "learning_rate": 0.0026029774183976206,
+ "loss": 4.982756614685059,
+ "step": 935
+ },
+ {
+ "epoch": 13.0,
+ "grad_norm": 0.011092593893408775,
+ "learning_rate": 0.002602069771090985,
+ "loss": 4.994824409484863,
+ "step": 936
+ },
+ {
+ "epoch": 13.0,
+ "eval_loss": 5.13364315032959,
+ "eval_runtime": 50.4195,
+ "eval_samples_per_second": 48.434,
+ "eval_steps_per_second": 1.527,
+ "step": 936
+ },
+ {
+ "epoch": 13.013973799126637,
+ "grad_norm": 0.01005905494093895,
+ "learning_rate": 0.0026011612479518694,
+ "loss": 4.918412685394287,
+ "step": 937
+ },
+ {
+ "epoch": 13.027947598253276,
+ "grad_norm": 0.009904682636260986,
+ "learning_rate": 0.002600251849712255,
+ "loss": 4.924630165100098,
+ "step": 938
+ },
+ {
+ "epoch": 13.041921397379912,
+ "grad_norm": 0.009543103165924549,
+ "learning_rate": 0.0025993415771048273,
+ "loss": 4.940309524536133,
+ "step": 939
+ },
+ {
+ "epoch": 13.055895196506551,
+ "grad_norm": 0.011031470261514187,
+ "learning_rate": 0.0025984304308629765,
+ "loss": 4.924609184265137,
+ "step": 940
+ },
+ {
+ "epoch": 13.069868995633188,
+ "grad_norm": 0.009511022828519344,
+ "learning_rate": 0.002597518411720796,
+ "loss": 4.9102325439453125,
+ "step": 941
+ },
+ {
+ "epoch": 13.083842794759825,
+ "grad_norm": 0.00886621605604887,
+ "learning_rate": 0.0025966055204130847,
+ "loss": 4.9236907958984375,
+ "step": 942
+ },
+ {
+ "epoch": 13.097816593886463,
+ "grad_norm": 0.010057039558887482,
+ "learning_rate": 0.002595691757675341,
+ "loss": 4.908633232116699,
+ "step": 943
+ },
+ {
+ "epoch": 13.1117903930131,
+ "grad_norm": 0.01098555512726307,
+ "learning_rate": 0.0025947771242437684,
+ "loss": 4.935927391052246,
+ "step": 944
+ },
+ {
+ "epoch": 13.125764192139737,
+ "grad_norm": 0.008800056762993336,
+ "learning_rate": 0.0025938616208552694,
+ "loss": 4.873291969299316,
+ "step": 945
+ },
+ {
+ "epoch": 13.139737991266376,
+ "grad_norm": 0.009747766889631748,
+ "learning_rate": 0.0025929452482474497,
+ "loss": 4.954436302185059,
+ "step": 946
+ },
+ {
+ "epoch": 13.153711790393013,
+ "grad_norm": 0.010785684920847416,
+ "learning_rate": 0.002592028007158613,
+ "loss": 4.924927711486816,
+ "step": 947
+ },
+ {
+ "epoch": 13.167685589519651,
+ "grad_norm": 0.012620717287063599,
+ "learning_rate": 0.0025911098983277645,
+ "loss": 4.953606605529785,
+ "step": 948
+ },
+ {
+ "epoch": 13.181659388646288,
+ "grad_norm": 0.014428738504648209,
+ "learning_rate": 0.002590190922494607,
+ "loss": 4.951706886291504,
+ "step": 949
+ },
+ {
+ "epoch": 13.195633187772925,
+ "grad_norm": 0.014496048912405968,
+ "learning_rate": 0.0025892710803995444,
+ "loss": 4.941249847412109,
+ "step": 950
+ },
+ {
+ "epoch": 13.209606986899564,
+ "grad_norm": 0.0109238987788558,
+ "learning_rate": 0.0025883503727836764,
+ "loss": 4.939975261688232,
+ "step": 951
+ },
+ {
+ "epoch": 13.2235807860262,
+ "grad_norm": 0.012271072715520859,
+ "learning_rate": 0.0025874288003887996,
+ "loss": 4.880942344665527,
+ "step": 952
+ },
+ {
+ "epoch": 13.237554585152838,
+ "grad_norm": 0.016302887350320816,
+ "learning_rate": 0.0025865063639574086,
+ "loss": 4.931602478027344,
+ "step": 953
+ },
+ {
+ "epoch": 13.251528384279476,
+ "grad_norm": 0.016137251630425453,
+ "learning_rate": 0.0025855830642326945,
+ "loss": 4.918013095855713,
+ "step": 954
+ },
+ {
+ "epoch": 13.265502183406113,
+ "grad_norm": 0.014732961542904377,
+ "learning_rate": 0.002584658901958542,
+ "loss": 4.927577495574951,
+ "step": 955
+ },
+ {
+ "epoch": 13.279475982532752,
+ "grad_norm": 0.016387946903705597,
+ "learning_rate": 0.0025837338778795325,
+ "loss": 4.930819034576416,
+ "step": 956
+ },
+ {
+ "epoch": 13.293449781659389,
+ "grad_norm": 0.017146151512861252,
+ "learning_rate": 0.002582807992740941,
+ "loss": 4.928377151489258,
+ "step": 957
+ },
+ {
+ "epoch": 13.307423580786025,
+ "grad_norm": 0.016153613105416298,
+ "learning_rate": 0.0025818812472887375,
+ "loss": 4.975607395172119,
+ "step": 958
+ },
+ {
+ "epoch": 13.321397379912664,
+ "grad_norm": 0.013386047445237637,
+ "learning_rate": 0.0025809536422695825,
+ "loss": 4.9434709548950195,
+ "step": 959
+ },
+ {
+ "epoch": 13.335371179039301,
+ "grad_norm": 0.012328253127634525,
+ "learning_rate": 0.0025800251784308314,
+ "loss": 4.917585372924805,
+ "step": 960
+ },
+ {
+ "epoch": 13.34934497816594,
+ "grad_norm": 0.012000586837530136,
+ "learning_rate": 0.0025790958565205316,
+ "loss": 4.9137725830078125,
+ "step": 961
+ },
+ {
+ "epoch": 13.363318777292577,
+ "grad_norm": 0.009463997557759285,
+ "learning_rate": 0.002578165677287419,
+ "loss": 4.910362243652344,
+ "step": 962
+ },
+ {
+ "epoch": 13.377292576419213,
+ "grad_norm": 0.010591360740363598,
+ "learning_rate": 0.0025772346414809242,
+ "loss": 4.916032314300537,
+ "step": 963
+ },
+ {
+ "epoch": 13.391266375545852,
+ "grad_norm": 0.010429762303829193,
+ "learning_rate": 0.002576302749851164,
+ "loss": 4.934561729431152,
+ "step": 964
+ },
+ {
+ "epoch": 13.405240174672489,
+ "grad_norm": 0.12526583671569824,
+ "learning_rate": 0.0025753700031489486,
+ "loss": 4.897515773773193,
+ "step": 965
+ },
+ {
+ "epoch": 13.419213973799126,
+ "grad_norm": 0.01368892751634121,
+ "learning_rate": 0.0025744364021257743,
+ "loss": 4.94814920425415,
+ "step": 966
+ },
+ {
+ "epoch": 13.433187772925764,
+ "grad_norm": 0.016603967174887657,
+ "learning_rate": 0.0025735019475338254,
+ "loss": 4.8996758460998535,
+ "step": 967
+ },
+ {
+ "epoch": 13.447161572052401,
+ "grad_norm": 0.06110106781125069,
+ "learning_rate": 0.0025725666401259775,
+ "loss": 4.946794509887695,
+ "step": 968
+ },
+ {
+ "epoch": 13.46113537117904,
+ "grad_norm": 0.018292423337697983,
+ "learning_rate": 0.002571630480655789,
+ "loss": 4.958992958068848,
+ "step": 969
+ },
+ {
+ "epoch": 13.475109170305677,
+ "grad_norm": 0.01852245442569256,
+ "learning_rate": 0.0025706934698775073,
+ "loss": 4.942556381225586,
+ "step": 970
+ },
+ {
+ "epoch": 13.489082969432314,
+ "grad_norm": 0.0195649154484272,
+ "learning_rate": 0.0025697556085460646,
+ "loss": 4.895773887634277,
+ "step": 971
+ },
+ {
+ "epoch": 13.503056768558952,
+ "grad_norm": 0.017820244655013084,
+ "learning_rate": 0.00256881689741708,
+ "loss": 4.915250778198242,
+ "step": 972
+ },
+ {
+ "epoch": 13.51703056768559,
+ "grad_norm": 0.016814742237329483,
+ "learning_rate": 0.002567877337246855,
+ "loss": 4.9521260261535645,
+ "step": 973
+ },
+ {
+ "epoch": 13.531004366812226,
+ "grad_norm": 0.015319781377911568,
+ "learning_rate": 0.002566936928792377,
+ "loss": 4.899402141571045,
+ "step": 974
+ },
+ {
+ "epoch": 13.544978165938865,
+ "grad_norm": 0.01607401855289936,
+ "learning_rate": 0.0025659956728113146,
+ "loss": 4.909913063049316,
+ "step": 975
+ },
+ {
+ "epoch": 13.558951965065502,
+ "grad_norm": 0.015564955770969391,
+ "learning_rate": 0.002565053570062023,
+ "loss": 4.947719573974609,
+ "step": 976
+ },
+ {
+ "epoch": 13.57292576419214,
+ "grad_norm": 0.017502037808299065,
+ "learning_rate": 0.002564110621303536,
+ "loss": 4.95107364654541,
+ "step": 977
+ },
+ {
+ "epoch": 13.586899563318777,
+ "grad_norm": 0.014791073277592659,
+ "learning_rate": 0.002563166827295571,
+ "loss": 4.8995866775512695,
+ "step": 978
+ },
+ {
+ "epoch": 13.600873362445414,
+ "grad_norm": 0.012700124643743038,
+ "learning_rate": 0.002562222188798525,
+ "loss": 4.922100067138672,
+ "step": 979
+ },
+ {
+ "epoch": 13.614847161572053,
+ "grad_norm": 0.011235272511839867,
+ "learning_rate": 0.0025612767065734763,
+ "loss": 4.884456634521484,
+ "step": 980
+ },
+ {
+ "epoch": 13.62882096069869,
+ "grad_norm": 0.013009768910706043,
+ "learning_rate": 0.002560330381382183,
+ "loss": 4.869503021240234,
+ "step": 981
+ },
+ {
+ "epoch": 13.642794759825328,
+ "grad_norm": 0.0102638378739357,
+ "learning_rate": 0.0025593832139870834,
+ "loss": 4.934551239013672,
+ "step": 982
+ },
+ {
+ "epoch": 13.656768558951965,
+ "grad_norm": 0.011081870645284653,
+ "learning_rate": 0.0025584352051512913,
+ "loss": 4.937268257141113,
+ "step": 983
+ },
+ {
+ "epoch": 13.670742358078602,
+ "grad_norm": 0.010681218467652798,
+ "learning_rate": 0.002557486355638601,
+ "loss": 4.920559883117676,
+ "step": 984
+ },
+ {
+ "epoch": 13.68471615720524,
+ "grad_norm": 0.010962890461087227,
+ "learning_rate": 0.002556536666213484,
+ "loss": 4.936082363128662,
+ "step": 985
+ },
+ {
+ "epoch": 13.698689956331878,
+ "grad_norm": 0.010790379717946053,
+ "learning_rate": 0.0025555861376410868,
+ "loss": 4.9506449699401855,
+ "step": 986
+ },
+ {
+ "epoch": 13.712663755458514,
+ "grad_norm": 0.010659157298505306,
+ "learning_rate": 0.0025546347706872346,
+ "loss": 4.872883319854736,
+ "step": 987
+ },
+ {
+ "epoch": 13.726637554585153,
+ "grad_norm": 0.012962655164301395,
+ "learning_rate": 0.0025536825661184254,
+ "loss": 4.901210784912109,
+ "step": 988
+ },
+ {
+ "epoch": 13.74061135371179,
+ "grad_norm": 0.014810293912887573,
+ "learning_rate": 0.002552729524701833,
+ "loss": 4.895804405212402,
+ "step": 989
+ },
+ {
+ "epoch": 13.754585152838429,
+ "grad_norm": 0.010993832722306252,
+ "learning_rate": 0.0025517756472053068,
+ "loss": 4.891035079956055,
+ "step": 990
+ },
+ {
+ "epoch": 13.768558951965066,
+ "grad_norm": 0.013413915410637856,
+ "learning_rate": 0.002550820934397368,
+ "loss": 4.900317192077637,
+ "step": 991
+ },
+ {
+ "epoch": 13.782532751091702,
+ "grad_norm": 0.01580147072672844,
+ "learning_rate": 0.0025498653870472117,
+ "loss": 4.883669853210449,
+ "step": 992
+ },
+ {
+ "epoch": 13.796506550218341,
+ "grad_norm": 0.01864512264728546,
+ "learning_rate": 0.0025489090059247045,
+ "loss": 4.881938457489014,
+ "step": 993
+ },
+ {
+ "epoch": 13.810480349344978,
+ "grad_norm": 0.017185743898153305,
+ "learning_rate": 0.0025479517918003862,
+ "loss": 4.853325843811035,
+ "step": 994
+ },
+ {
+ "epoch": 13.824454148471617,
+ "grad_norm": 0.012045156210660934,
+ "learning_rate": 0.002546993745445466,
+ "loss": 4.9152326583862305,
+ "step": 995
+ },
+ {
+ "epoch": 13.838427947598253,
+ "grad_norm": 0.013420728035271168,
+ "learning_rate": 0.0025460348676318257,
+ "loss": 4.895130157470703,
+ "step": 996
+ },
+ {
+ "epoch": 13.85240174672489,
+ "grad_norm": 0.011561610735952854,
+ "learning_rate": 0.0025450751591320152,
+ "loss": 4.890257835388184,
+ "step": 997
+ },
+ {
+ "epoch": 13.866375545851529,
+ "grad_norm": 0.013164008036255836,
+ "learning_rate": 0.0025441146207192547,
+ "loss": 4.885972023010254,
+ "step": 998
+ },
+ {
+ "epoch": 13.880349344978166,
+ "grad_norm": 0.01120204571634531,
+ "learning_rate": 0.002543153253167432,
+ "loss": 4.882246971130371,
+ "step": 999
+ },
+ {
+ "epoch": 13.894323144104803,
+ "grad_norm": 0.011342627927660942,
+ "learning_rate": 0.002542191057251104,
+ "loss": 4.862927436828613,
+ "step": 1000
+ },
+ {
+ "epoch": 13.908296943231441,
+ "grad_norm": 0.010852213948965073,
+ "learning_rate": 0.0025412280337454942,
+ "loss": 4.879744529724121,
+ "step": 1001
+ },
+ {
+ "epoch": 13.922270742358078,
+ "grad_norm": 0.012851282022893429,
+ "learning_rate": 0.0025402641834264936,
+ "loss": 4.882238388061523,
+ "step": 1002
+ },
+ {
+ "epoch": 13.936244541484717,
+ "grad_norm": 0.011839655227959156,
+ "learning_rate": 0.002539299507070659,
+ "loss": 4.85658073425293,
+ "step": 1003
+ },
+ {
+ "epoch": 13.950218340611354,
+ "grad_norm": 0.011315262876451015,
+ "learning_rate": 0.002538334005455213,
+ "loss": 4.9234771728515625,
+ "step": 1004
+ },
+ {
+ "epoch": 13.96419213973799,
+ "grad_norm": 0.011661105789244175,
+ "learning_rate": 0.0025373676793580416,
+ "loss": 4.855696678161621,
+ "step": 1005
+ },
+ {
+ "epoch": 13.97816593886463,
+ "grad_norm": 0.009712521918118,
+ "learning_rate": 0.0025364005295576975,
+ "loss": 4.89942741394043,
+ "step": 1006
+ },
+ {
+ "epoch": 13.992139737991266,
+ "grad_norm": 0.011002928018569946,
+ "learning_rate": 0.0025354325568333956,
+ "loss": 4.871459007263184,
+ "step": 1007
+ },
+ {
+ "epoch": 14.0,
+ "grad_norm": 0.01509513333439827,
+ "learning_rate": 0.0025344637619650135,
+ "loss": 4.834778785705566,
+ "step": 1008
+ },
+ {
+ "epoch": 14.0,
+ "eval_loss": 5.031810283660889,
+ "eval_runtime": 51.9412,
+ "eval_samples_per_second": 47.015,
+ "eval_steps_per_second": 1.482,
+ "step": 1008
+ },
+ {
+ "epoch": 14.013973799126637,
+ "grad_norm": 0.010317740961909294,
+ "learning_rate": 0.0025334941457330925,
+ "loss": 4.8169708251953125,
+ "step": 1009
+ },
+ {
+ "epoch": 14.027947598253276,
+ "grad_norm": 0.00892715621739626,
+ "learning_rate": 0.0025325237089188343,
+ "loss": 4.852178573608398,
+ "step": 1010
+ },
+ {
+ "epoch": 14.041921397379912,
+ "grad_norm": 0.011608307249844074,
+ "learning_rate": 0.002531552452304103,
+ "loss": 4.790487289428711,
+ "step": 1011
+ },
+ {
+ "epoch": 14.055895196506551,
+ "grad_norm": 0.013100114651024342,
+ "learning_rate": 0.0025305803766714216,
+ "loss": 4.863536834716797,
+ "step": 1012
+ },
+ {
+ "epoch": 14.069868995633188,
+ "grad_norm": 0.016571057960391045,
+ "learning_rate": 0.002529607482803974,
+ "loss": 4.840987205505371,
+ "step": 1013
+ },
+ {
+ "epoch": 14.083842794759825,
+ "grad_norm": 0.014595622196793556,
+ "learning_rate": 0.002528633771485604,
+ "loss": 4.815027236938477,
+ "step": 1014
+ },
+ {
+ "epoch": 14.097816593886463,
+ "grad_norm": 0.014603931456804276,
+ "learning_rate": 0.002527659243500812,
+ "loss": 4.853024959564209,
+ "step": 1015
+ },
+ {
+ "epoch": 14.1117903930131,
+ "grad_norm": 0.019107041880488396,
+ "learning_rate": 0.0025266838996347595,
+ "loss": 4.829808235168457,
+ "step": 1016
+ },
+ {
+ "epoch": 14.125764192139737,
+ "grad_norm": 0.020292190834879875,
+ "learning_rate": 0.0025257077406732613,
+ "loss": 4.775232315063477,
+ "step": 1017
+ },
+ {
+ "epoch": 14.139737991266376,
+ "grad_norm": 0.014630074612796307,
+ "learning_rate": 0.0025247307674027926,
+ "loss": 4.826682090759277,
+ "step": 1018
+ },
+ {
+ "epoch": 14.153711790393013,
+ "grad_norm": 0.019507668912410736,
+ "learning_rate": 0.0025237529806104823,
+ "loss": 4.843222141265869,
+ "step": 1019
+ },
+ {
+ "epoch": 14.167685589519651,
+ "grad_norm": 0.023950330913066864,
+ "learning_rate": 0.002522774381084116,
+ "loss": 4.820512771606445,
+ "step": 1020
+ },
+ {
+ "epoch": 14.181659388646288,
+ "grad_norm": 0.022547820582985878,
+ "learning_rate": 0.002521794969612133,
+ "loss": 4.854910850524902,
+ "step": 1021
+ },
+ {
+ "epoch": 14.195633187772925,
+ "grad_norm": 0.017819277942180634,
+ "learning_rate": 0.0025208147469836283,
+ "loss": 4.825885772705078,
+ "step": 1022
+ },
+ {
+ "epoch": 14.209606986899564,
+ "grad_norm": 0.017989592626690865,
+ "learning_rate": 0.002519833713988349,
+ "loss": 4.8357014656066895,
+ "step": 1023
+ },
+ {
+ "epoch": 14.2235807860262,
+ "grad_norm": 0.013516640290617943,
+ "learning_rate": 0.0025188518714166956,
+ "loss": 4.83729362487793,
+ "step": 1024
+ },
+ {
+ "epoch": 14.237554585152838,
+ "grad_norm": 0.015956101939082146,
+ "learning_rate": 0.002517869220059721,
+ "loss": 4.807845592498779,
+ "step": 1025
+ },
+ {
+ "epoch": 14.251528384279476,
+ "grad_norm": 0.013967473991215229,
+ "learning_rate": 0.002516885760709131,
+ "loss": 4.828944206237793,
+ "step": 1026
+ },
+ {
+ "epoch": 14.265502183406113,
+ "grad_norm": 0.01270931027829647,
+ "learning_rate": 0.0025159014941572785,
+ "loss": 4.826594352722168,
+ "step": 1027
+ },
+ {
+ "epoch": 14.279475982532752,
+ "grad_norm": 0.012433716095983982,
+ "learning_rate": 0.002514916421197171,
+ "loss": 4.797939777374268,
+ "step": 1028
+ },
+ {
+ "epoch": 14.293449781659389,
+ "grad_norm": 0.011882808059453964,
+ "learning_rate": 0.0025139305426224636,
+ "loss": 4.78917932510376,
+ "step": 1029
+ },
+ {
+ "epoch": 14.307423580786025,
+ "grad_norm": 0.011193005368113518,
+ "learning_rate": 0.0025129438592274614,
+ "loss": 4.787964820861816,
+ "step": 1030
+ },
+ {
+ "epoch": 14.321397379912664,
+ "grad_norm": 0.011873407289385796,
+ "learning_rate": 0.002511956371807116,
+ "loss": 4.81754732131958,
+ "step": 1031
+ },
+ {
+ "epoch": 14.335371179039301,
+ "grad_norm": 0.011407418176531792,
+ "learning_rate": 0.002510968081157029,
+ "loss": 4.781639575958252,
+ "step": 1032
+ },
+ {
+ "epoch": 14.34934497816594,
+ "grad_norm": 0.009560304693877697,
+ "learning_rate": 0.002509978988073449,
+ "loss": 4.8163957595825195,
+ "step": 1033
+ },
+ {
+ "epoch": 14.363318777292577,
+ "grad_norm": 0.009969786740839481,
+ "learning_rate": 0.0025089890933532697,
+ "loss": 4.795045852661133,
+ "step": 1034
+ },
+ {
+ "epoch": 14.377292576419213,
+ "grad_norm": 0.010739767923951149,
+ "learning_rate": 0.0025079983977940312,
+ "loss": 4.847584247589111,
+ "step": 1035
+ },
+ {
+ "epoch": 14.391266375545852,
+ "grad_norm": 0.009999158792197704,
+ "learning_rate": 0.0025070069021939194,
+ "loss": 4.834909439086914,
+ "step": 1036
+ },
+ {
+ "epoch": 14.405240174672489,
+ "grad_norm": 0.009302482940256596,
+ "learning_rate": 0.002506014607351765,
+ "loss": 4.808745384216309,
+ "step": 1037
+ },
+ {
+ "epoch": 14.419213973799126,
+ "grad_norm": 0.010687744244933128,
+ "learning_rate": 0.0025050215140670403,
+ "loss": 4.789101600646973,
+ "step": 1038
+ },
+ {
+ "epoch": 14.433187772925764,
+ "grad_norm": 0.010930616408586502,
+ "learning_rate": 0.0025040276231398637,
+ "loss": 4.817464351654053,
+ "step": 1039
+ },
+ {
+ "epoch": 14.447161572052401,
+ "grad_norm": 0.013101065531373024,
+ "learning_rate": 0.0025030329353709956,
+ "loss": 4.827780723571777,
+ "step": 1040
+ },
+ {
+ "epoch": 14.46113537117904,
+ "grad_norm": 0.01223307941108942,
+ "learning_rate": 0.0025020374515618373,
+ "loss": 4.8367414474487305,
+ "step": 1041
+ },
+ {
+ "epoch": 14.475109170305677,
+ "grad_norm": 0.013091634958982468,
+ "learning_rate": 0.0025010411725144323,
+ "loss": 4.839454650878906,
+ "step": 1042
+ },
+ {
+ "epoch": 14.489082969432314,
+ "grad_norm": 0.012114843353629112,
+ "learning_rate": 0.0025000440990314643,
+ "loss": 4.832447052001953,
+ "step": 1043
+ },
+ {
+ "epoch": 14.503056768558952,
+ "grad_norm": 0.014664588496088982,
+ "learning_rate": 0.0024990462319162588,
+ "loss": 4.740812301635742,
+ "step": 1044
+ },
+ {
+ "epoch": 14.51703056768559,
+ "grad_norm": 0.015566283836960793,
+ "learning_rate": 0.002498047571972777,
+ "loss": 4.815671920776367,
+ "step": 1045
+ },
+ {
+ "epoch": 14.531004366812226,
+ "grad_norm": 0.012440434657037258,
+ "learning_rate": 0.002497048120005623,
+ "loss": 4.769775867462158,
+ "step": 1046
+ },
+ {
+ "epoch": 14.544978165938865,
+ "grad_norm": 0.01175499614328146,
+ "learning_rate": 0.0024960478768200362,
+ "loss": 4.8101348876953125,
+ "step": 1047
+ },
+ {
+ "epoch": 14.558951965065502,
+ "grad_norm": 0.012699508108198643,
+ "learning_rate": 0.002495046843221896,
+ "loss": 4.827281951904297,
+ "step": 1048
+ },
+ {
+ "epoch": 14.57292576419214,
+ "grad_norm": 0.012438303790986538,
+ "learning_rate": 0.002494045020017715,
+ "loss": 4.786289691925049,
+ "step": 1049
+ },
+ {
+ "epoch": 14.586899563318777,
+ "grad_norm": 0.01709924265742302,
+ "learning_rate": 0.002493042408014646,
+ "loss": 4.828742027282715,
+ "step": 1050
+ },
+ {
+ "epoch": 14.600873362445414,
+ "grad_norm": 0.022684535011649132,
+ "learning_rate": 0.0024920390080204733,
+ "loss": 4.822258949279785,
+ "step": 1051
+ },
+ {
+ "epoch": 14.614847161572053,
+ "grad_norm": 0.011984334327280521,
+ "learning_rate": 0.00249103482084362,
+ "loss": 4.792006969451904,
+ "step": 1052
+ },
+ {
+ "epoch": 14.62882096069869,
+ "grad_norm": 0.013941102661192417,
+ "learning_rate": 0.00249002984729314,
+ "loss": 4.7926177978515625,
+ "step": 1053
+ },
+ {
+ "epoch": 14.642794759825328,
+ "grad_norm": 0.0145634226500988,
+ "learning_rate": 0.002489024088178724,
+ "loss": 4.819155216217041,
+ "step": 1054
+ },
+ {
+ "epoch": 14.656768558951965,
+ "grad_norm": 0.013391612097620964,
+ "learning_rate": 0.0024880175443106932,
+ "loss": 4.777242660522461,
+ "step": 1055
+ },
+ {
+ "epoch": 14.670742358078602,
+ "grad_norm": 0.01764756441116333,
+ "learning_rate": 0.0024870102165000015,
+ "loss": 4.81627893447876,
+ "step": 1056
+ },
+ {
+ "epoch": 14.68471615720524,
+ "grad_norm": 0.015397441573441029,
+ "learning_rate": 0.002486002105558235,
+ "loss": 4.869495391845703,
+ "step": 1057
+ },
+ {
+ "epoch": 14.698689956331878,
+ "grad_norm": 0.01225620973855257,
+ "learning_rate": 0.0024849932122976105,
+ "loss": 4.784395217895508,
+ "step": 1058
+ },
+ {
+ "epoch": 14.712663755458514,
+ "grad_norm": 0.01534759160131216,
+ "learning_rate": 0.002483983537530975,
+ "loss": 4.813475608825684,
+ "step": 1059
+ },
+ {
+ "epoch": 14.726637554585153,
+ "grad_norm": 0.015057779848575592,
+ "learning_rate": 0.002482973082071805,
+ "loss": 4.8281569480896,
+ "step": 1060
+ },
+ {
+ "epoch": 14.74061135371179,
+ "grad_norm": 0.01243986189365387,
+ "learning_rate": 0.0024819618467342066,
+ "loss": 4.7866973876953125,
+ "step": 1061
+ },
+ {
+ "epoch": 14.754585152838429,
+ "grad_norm": 0.013326527550816536,
+ "learning_rate": 0.0024809498323329136,
+ "loss": 4.750214576721191,
+ "step": 1062
+ },
+ {
+ "epoch": 14.768558951965066,
+ "grad_norm": 0.013644695281982422,
+ "learning_rate": 0.002479937039683288,
+ "loss": 4.790144920349121,
+ "step": 1063
+ },
+ {
+ "epoch": 14.782532751091702,
+ "grad_norm": 0.01296213548630476,
+ "learning_rate": 0.002478923469601319,
+ "loss": 4.812170028686523,
+ "step": 1064
+ },
+ {
+ "epoch": 14.796506550218341,
+ "grad_norm": 0.011848201975226402,
+ "learning_rate": 0.0024779091229036204,
+ "loss": 4.798855781555176,
+ "step": 1065
+ },
+ {
+ "epoch": 14.810480349344978,
+ "grad_norm": 0.012765968218445778,
+ "learning_rate": 0.0024768940004074335,
+ "loss": 4.8272294998168945,
+ "step": 1066
+ },
+ {
+ "epoch": 14.824454148471617,
+ "grad_norm": 0.011700666509568691,
+ "learning_rate": 0.002475878102930625,
+ "loss": 4.762265205383301,
+ "step": 1067
+ },
+ {
+ "epoch": 14.838427947598253,
+ "grad_norm": 0.01018498931080103,
+ "learning_rate": 0.0024748614312916843,
+ "loss": 4.796628952026367,
+ "step": 1068
+ },
+ {
+ "epoch": 14.85240174672489,
+ "grad_norm": 0.010241681709885597,
+ "learning_rate": 0.002473843986309725,
+ "loss": 4.829319000244141,
+ "step": 1069
+ },
+ {
+ "epoch": 14.866375545851529,
+ "grad_norm": 0.009332080371677876,
+ "learning_rate": 0.0024728257688044855,
+ "loss": 4.783773899078369,
+ "step": 1070
+ },
+ {
+ "epoch": 14.880349344978166,
+ "grad_norm": 0.012860726565122604,
+ "learning_rate": 0.002471806779596324,
+ "loss": 4.802700519561768,
+ "step": 1071
+ },
+ {
+ "epoch": 14.894323144104803,
+ "grad_norm": 0.011156697757542133,
+ "learning_rate": 0.0024707870195062225,
+ "loss": 4.808038234710693,
+ "step": 1072
+ },
+ {
+ "epoch": 14.908296943231441,
+ "grad_norm": 0.010329720564186573,
+ "learning_rate": 0.002469766489355783,
+ "loss": 4.761645317077637,
+ "step": 1073
+ },
+ {
+ "epoch": 14.922270742358078,
+ "grad_norm": 0.011093895882368088,
+ "learning_rate": 0.0024687451899672278,
+ "loss": 4.795220851898193,
+ "step": 1074
+ },
+ {
+ "epoch": 14.936244541484717,
+ "grad_norm": 0.010740782134234905,
+ "learning_rate": 0.0024677231221634,
+ "loss": 4.79289436340332,
+ "step": 1075
+ },
+ {
+ "epoch": 14.950218340611354,
+ "grad_norm": 0.012272508814930916,
+ "learning_rate": 0.0024667002867677606,
+ "loss": 4.811873435974121,
+ "step": 1076
+ },
+ {
+ "epoch": 14.96419213973799,
+ "grad_norm": 0.008665879257023335,
+ "learning_rate": 0.0024656766846043905,
+ "loss": 4.744179725646973,
+ "step": 1077
+ },
+ {
+ "epoch": 14.97816593886463,
+ "grad_norm": 0.00933373998850584,
+ "learning_rate": 0.0024646523164979866,
+ "loss": 4.805453777313232,
+ "step": 1078
+ },
+ {
+ "epoch": 14.992139737991266,
+ "grad_norm": 0.010399337857961655,
+ "learning_rate": 0.002463627183273864,
+ "loss": 4.75568962097168,
+ "step": 1079
+ },
+ {
+ "epoch": 15.0,
+ "grad_norm": 0.013210986740887165,
+ "learning_rate": 0.0024626012857579542,
+ "loss": 4.797974586486816,
+ "step": 1080
+ },
+ {
+ "epoch": 15.0,
+ "eval_loss": 4.967026233673096,
+ "eval_runtime": 51.07,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 1080
+ },
+ {
+ "epoch": 15.013973799126637,
+ "grad_norm": 0.01731858029961586,
+ "learning_rate": 0.0024615746247768043,
+ "loss": 4.748838901519775,
+ "step": 1081
+ },
+ {
+ "epoch": 15.027947598253276,
+ "grad_norm": 0.018902689218521118,
+ "learning_rate": 0.0024605472011575763,
+ "loss": 4.717338562011719,
+ "step": 1082
+ },
+ {
+ "epoch": 15.041921397379912,
+ "grad_norm": 0.01734415628015995,
+ "learning_rate": 0.0024595190157280468,
+ "loss": 4.737737655639648,
+ "step": 1083
+ },
+ {
+ "epoch": 15.055895196506551,
+ "grad_norm": 0.01971321739256382,
+ "learning_rate": 0.002458490069316607,
+ "loss": 4.774558067321777,
+ "step": 1084
+ },
+ {
+ "epoch": 15.069868995633188,
+ "grad_norm": 0.020029844716191292,
+ "learning_rate": 0.002457460362752259,
+ "loss": 4.739579200744629,
+ "step": 1085
+ },
+ {
+ "epoch": 15.083842794759825,
+ "grad_norm": 0.020853109657764435,
+ "learning_rate": 0.0024564298968646203,
+ "loss": 4.791436672210693,
+ "step": 1086
+ },
+ {
+ "epoch": 15.097816593886463,
+ "grad_norm": 0.019159914925694466,
+ "learning_rate": 0.0024553986724839176,
+ "loss": 4.765683650970459,
+ "step": 1087
+ },
+ {
+ "epoch": 15.1117903930131,
+ "grad_norm": 0.0165473110973835,
+ "learning_rate": 0.0024543666904409903,
+ "loss": 4.732603073120117,
+ "step": 1088
+ },
+ {
+ "epoch": 15.125764192139737,
+ "grad_norm": 0.017426826059818268,
+ "learning_rate": 0.002453333951567287,
+ "loss": 4.7360639572143555,
+ "step": 1089
+ },
+ {
+ "epoch": 15.139737991266376,
+ "grad_norm": 0.014590734615921974,
+ "learning_rate": 0.002452300456694868,
+ "loss": 4.750031471252441,
+ "step": 1090
+ },
+ {
+ "epoch": 15.153711790393013,
+ "grad_norm": 0.011172222904860973,
+ "learning_rate": 0.0024512662066563996,
+ "loss": 4.740771293640137,
+ "step": 1091
+ },
+ {
+ "epoch": 15.167685589519651,
+ "grad_norm": 0.012294141575694084,
+ "learning_rate": 0.0024502312022851606,
+ "loss": 4.726214408874512,
+ "step": 1092
+ },
+ {
+ "epoch": 15.181659388646288,
+ "grad_norm": 0.01146282535046339,
+ "learning_rate": 0.002449195444415034,
+ "loss": 4.753599166870117,
+ "step": 1093
+ },
+ {
+ "epoch": 15.195633187772925,
+ "grad_norm": 0.011197247542440891,
+ "learning_rate": 0.0024481589338805113,
+ "loss": 4.751479148864746,
+ "step": 1094
+ },
+ {
+ "epoch": 15.209606986899564,
+ "grad_norm": 0.013252820819616318,
+ "learning_rate": 0.0024471216715166904,
+ "loss": 4.743409156799316,
+ "step": 1095
+ },
+ {
+ "epoch": 15.2235807860262,
+ "grad_norm": 0.012366211041808128,
+ "learning_rate": 0.0024460836581592744,
+ "loss": 4.731625556945801,
+ "step": 1096
+ },
+ {
+ "epoch": 15.237554585152838,
+ "grad_norm": 0.013906655833125114,
+ "learning_rate": 0.0024450448946445737,
+ "loss": 4.768153190612793,
+ "step": 1097
+ },
+ {
+ "epoch": 15.251528384279476,
+ "grad_norm": 0.012118754908442497,
+ "learning_rate": 0.002444005381809499,
+ "loss": 4.74198579788208,
+ "step": 1098
+ },
+ {
+ "epoch": 15.265502183406113,
+ "grad_norm": 0.01197550818324089,
+ "learning_rate": 0.002442965120491568,
+ "loss": 4.737289905548096,
+ "step": 1099
+ },
+ {
+ "epoch": 15.279475982532752,
+ "grad_norm": 0.010640372522175312,
+ "learning_rate": 0.002441924111528901,
+ "loss": 4.7472686767578125,
+ "step": 1100
+ },
+ {
+ "epoch": 15.293449781659389,
+ "grad_norm": 0.013553396798670292,
+ "learning_rate": 0.0024408823557602196,
+ "loss": 4.711215972900391,
+ "step": 1101
+ },
+ {
+ "epoch": 15.307423580786025,
+ "grad_norm": 0.01393711194396019,
+ "learning_rate": 0.002439839854024847,
+ "loss": 4.720060348510742,
+ "step": 1102
+ },
+ {
+ "epoch": 15.321397379912664,
+ "grad_norm": 0.01589774526655674,
+ "learning_rate": 0.002438796607162709,
+ "loss": 4.695697784423828,
+ "step": 1103
+ },
+ {
+ "epoch": 15.335371179039301,
+ "grad_norm": 0.0097908740863204,
+ "learning_rate": 0.0024377526160143303,
+ "loss": 4.729333877563477,
+ "step": 1104
+ },
+ {
+ "epoch": 15.34934497816594,
+ "grad_norm": 0.011745489202439785,
+ "learning_rate": 0.0024367078814208363,
+ "loss": 4.722957134246826,
+ "step": 1105
+ },
+ {
+ "epoch": 15.363318777292577,
+ "grad_norm": 0.009649662300944328,
+ "learning_rate": 0.0024356624042239498,
+ "loss": 4.727540493011475,
+ "step": 1106
+ },
+ {
+ "epoch": 15.377292576419213,
+ "grad_norm": 0.008429240435361862,
+ "learning_rate": 0.002434616185265993,
+ "loss": 4.705052375793457,
+ "step": 1107
+ },
+ {
+ "epoch": 15.391266375545852,
+ "grad_norm": 0.00984937883913517,
+ "learning_rate": 0.002433569225389887,
+ "loss": 4.721161365509033,
+ "step": 1108
+ },
+ {
+ "epoch": 15.405240174672489,
+ "grad_norm": 0.009237290360033512,
+ "learning_rate": 0.0024325215254391466,
+ "loss": 4.724277496337891,
+ "step": 1109
+ },
+ {
+ "epoch": 15.419213973799126,
+ "grad_norm": 0.008900545537471771,
+ "learning_rate": 0.002431473086257886,
+ "loss": 4.711522102355957,
+ "step": 1110
+ },
+ {
+ "epoch": 15.433187772925764,
+ "grad_norm": 0.010003026574850082,
+ "learning_rate": 0.002430423908690813,
+ "loss": 4.728816032409668,
+ "step": 1111
+ },
+ {
+ "epoch": 15.447161572052401,
+ "grad_norm": 0.011658324860036373,
+ "learning_rate": 0.0024293739935832316,
+ "loss": 4.716179847717285,
+ "step": 1112
+ },
+ {
+ "epoch": 15.46113537117904,
+ "grad_norm": 0.01312185451388359,
+ "learning_rate": 0.0024283233417810393,
+ "loss": 4.7172346115112305,
+ "step": 1113
+ },
+ {
+ "epoch": 15.475109170305677,
+ "grad_norm": 0.014306646771728992,
+ "learning_rate": 0.0024272719541307266,
+ "loss": 4.704054832458496,
+ "step": 1114
+ },
+ {
+ "epoch": 15.489082969432314,
+ "grad_norm": 0.009844634681940079,
+ "learning_rate": 0.0024262198314793783,
+ "loss": 4.72379207611084,
+ "step": 1115
+ },
+ {
+ "epoch": 15.503056768558952,
+ "grad_norm": 0.00903896801173687,
+ "learning_rate": 0.0024251669746746703,
+ "loss": 4.711025238037109,
+ "step": 1116
+ },
+ {
+ "epoch": 15.51703056768559,
+ "grad_norm": 0.010476433672010899,
+ "learning_rate": 0.0024241133845648703,
+ "loss": 4.711664199829102,
+ "step": 1117
+ },
+ {
+ "epoch": 15.531004366812226,
+ "grad_norm": 0.009128391742706299,
+ "learning_rate": 0.0024230590619988375,
+ "loss": 4.72395133972168,
+ "step": 1118
+ },
+ {
+ "epoch": 15.544978165938865,
+ "grad_norm": 0.01228832732886076,
+ "learning_rate": 0.0024220040078260195,
+ "loss": 4.747347831726074,
+ "step": 1119
+ },
+ {
+ "epoch": 15.558951965065502,
+ "grad_norm": 0.011916602961719036,
+ "learning_rate": 0.002420948222896455,
+ "loss": 4.696471691131592,
+ "step": 1120
+ },
+ {
+ "epoch": 15.57292576419214,
+ "grad_norm": 0.012877700850367546,
+ "learning_rate": 0.0024198917080607706,
+ "loss": 4.731119632720947,
+ "step": 1121
+ },
+ {
+ "epoch": 15.586899563318777,
+ "grad_norm": 0.012967764399945736,
+ "learning_rate": 0.0024188344641701814,
+ "loss": 4.688953876495361,
+ "step": 1122
+ },
+ {
+ "epoch": 15.600873362445414,
+ "grad_norm": 0.011993546038866043,
+ "learning_rate": 0.002417776492076489,
+ "loss": 4.695237159729004,
+ "step": 1123
+ },
+ {
+ "epoch": 15.614847161572053,
+ "grad_norm": 0.011188571341335773,
+ "learning_rate": 0.0024167177926320833,
+ "loss": 4.699850082397461,
+ "step": 1124
+ },
+ {
+ "epoch": 15.62882096069869,
+ "grad_norm": 0.010556348599493504,
+ "learning_rate": 0.0024156583666899395,
+ "loss": 4.703063011169434,
+ "step": 1125
+ },
+ {
+ "epoch": 15.642794759825328,
+ "grad_norm": 0.012023619376122952,
+ "learning_rate": 0.0024145982151036176,
+ "loss": 4.6918792724609375,
+ "step": 1126
+ },
+ {
+ "epoch": 15.656768558951965,
+ "grad_norm": 0.01266001258045435,
+ "learning_rate": 0.0024135373387272623,
+ "loss": 4.6653289794921875,
+ "step": 1127
+ },
+ {
+ "epoch": 15.670742358078602,
+ "grad_norm": 0.012152725830674171,
+ "learning_rate": 0.0024124757384156024,
+ "loss": 4.682897567749023,
+ "step": 1128
+ },
+ {
+ "epoch": 15.68471615720524,
+ "grad_norm": 0.014056944288313389,
+ "learning_rate": 0.002411413415023951,
+ "loss": 4.696136474609375,
+ "step": 1129
+ },
+ {
+ "epoch": 15.698689956331878,
+ "grad_norm": 0.016826556995511055,
+ "learning_rate": 0.002410350369408202,
+ "loss": 4.701657295227051,
+ "step": 1130
+ },
+ {
+ "epoch": 15.712663755458514,
+ "grad_norm": 0.015048800967633724,
+ "learning_rate": 0.0024092866024248323,
+ "loss": 4.691389083862305,
+ "step": 1131
+ },
+ {
+ "epoch": 15.726637554585153,
+ "grad_norm": 0.012512968853116035,
+ "learning_rate": 0.0024082221149308997,
+ "loss": 4.677886009216309,
+ "step": 1132
+ },
+ {
+ "epoch": 15.74061135371179,
+ "grad_norm": 0.01596425101161003,
+ "learning_rate": 0.0024071569077840425,
+ "loss": 4.718337535858154,
+ "step": 1133
+ },
+ {
+ "epoch": 15.754585152838429,
+ "grad_norm": 0.01640196517109871,
+ "learning_rate": 0.002406090981842479,
+ "loss": 4.697303295135498,
+ "step": 1134
+ },
+ {
+ "epoch": 15.768558951965066,
+ "grad_norm": 0.012347170151770115,
+ "learning_rate": 0.002405024337965007,
+ "loss": 4.696938991546631,
+ "step": 1135
+ },
+ {
+ "epoch": 15.782532751091702,
+ "grad_norm": 0.01213028747588396,
+ "learning_rate": 0.0024039569770110012,
+ "loss": 4.737306594848633,
+ "step": 1136
+ },
+ {
+ "epoch": 15.796506550218341,
+ "grad_norm": 0.01314542256295681,
+ "learning_rate": 0.002402888899840416,
+ "loss": 4.69948673248291,
+ "step": 1137
+ },
+ {
+ "epoch": 15.810480349344978,
+ "grad_norm": 0.013764189556241035,
+ "learning_rate": 0.0024018201073137805,
+ "loss": 4.722800254821777,
+ "step": 1138
+ },
+ {
+ "epoch": 15.824454148471617,
+ "grad_norm": 0.011147351004183292,
+ "learning_rate": 0.002400750600292203,
+ "loss": 4.694892883300781,
+ "step": 1139
+ },
+ {
+ "epoch": 15.838427947598253,
+ "grad_norm": 0.015437614172697067,
+ "learning_rate": 0.0023996803796373653,
+ "loss": 4.704752445220947,
+ "step": 1140
+ },
+ {
+ "epoch": 15.85240174672489,
+ "grad_norm": 0.016402097418904305,
+ "learning_rate": 0.0023986094462115255,
+ "loss": 4.67849063873291,
+ "step": 1141
+ },
+ {
+ "epoch": 15.866375545851529,
+ "grad_norm": 0.011600594967603683,
+ "learning_rate": 0.002397537800877514,
+ "loss": 4.689641952514648,
+ "step": 1142
+ },
+ {
+ "epoch": 15.880349344978166,
+ "grad_norm": 0.011864298023283482,
+ "learning_rate": 0.0023964654444987373,
+ "loss": 4.670377731323242,
+ "step": 1143
+ },
+ {
+ "epoch": 15.894323144104803,
+ "grad_norm": 0.01409872155636549,
+ "learning_rate": 0.002395392377939173,
+ "loss": 4.696022033691406,
+ "step": 1144
+ },
+ {
+ "epoch": 15.908296943231441,
+ "grad_norm": 0.012106554582715034,
+ "learning_rate": 0.0023943186020633717,
+ "loss": 4.705615043640137,
+ "step": 1145
+ },
+ {
+ "epoch": 15.922270742358078,
+ "grad_norm": 0.012310667894780636,
+ "learning_rate": 0.002393244117736455,
+ "loss": 4.702986717224121,
+ "step": 1146
+ },
+ {
+ "epoch": 15.936244541484717,
+ "grad_norm": 0.013501073233783245,
+ "learning_rate": 0.002392168925824116,
+ "loss": 4.6700215339660645,
+ "step": 1147
+ },
+ {
+ "epoch": 15.950218340611354,
+ "grad_norm": 0.013565126806497574,
+ "learning_rate": 0.0023910930271926166,
+ "loss": 4.699914455413818,
+ "step": 1148
+ },
+ {
+ "epoch": 15.96419213973799,
+ "grad_norm": 0.013421077281236649,
+ "learning_rate": 0.00239001642270879,
+ "loss": 4.713191986083984,
+ "step": 1149
+ },
+ {
+ "epoch": 15.97816593886463,
+ "grad_norm": 0.014408102259039879,
+ "learning_rate": 0.002388939113240035,
+ "loss": 4.655015468597412,
+ "step": 1150
+ },
+ {
+ "epoch": 15.992139737991266,
+ "grad_norm": 0.013502835296094418,
+ "learning_rate": 0.0023878610996543224,
+ "loss": 4.6142258644104,
+ "step": 1151
+ },
+ {
+ "epoch": 16.0,
+ "grad_norm": 0.014169570058584213,
+ "learning_rate": 0.002386782382820188,
+ "loss": 4.660243034362793,
+ "step": 1152
+ },
+ {
+ "epoch": 16.0,
+ "eval_loss": 4.885881423950195,
+ "eval_runtime": 52.0315,
+ "eval_samples_per_second": 46.933,
+ "eval_steps_per_second": 1.48,
+ "step": 1152
+ },
+ {
+ "epoch": 16.01397379912664,
+ "grad_norm": 0.014618083834648132,
+ "learning_rate": 0.002385702963606734,
+ "loss": 4.640046119689941,
+ "step": 1153
+ },
+ {
+ "epoch": 16.027947598253274,
+ "grad_norm": 0.01745622046291828,
+ "learning_rate": 0.0023846228428836278,
+ "loss": 4.641966819763184,
+ "step": 1154
+ },
+ {
+ "epoch": 16.041921397379912,
+ "grad_norm": 0.014995258301496506,
+ "learning_rate": 0.002383542021521106,
+ "loss": 4.60922908782959,
+ "step": 1155
+ },
+ {
+ "epoch": 16.05589519650655,
+ "grad_norm": 0.013278020545840263,
+ "learning_rate": 0.002382460500389964,
+ "loss": 4.652071952819824,
+ "step": 1156
+ },
+ {
+ "epoch": 16.069868995633186,
+ "grad_norm": 0.014685257337987423,
+ "learning_rate": 0.002381378280361566,
+ "loss": 4.672174453735352,
+ "step": 1157
+ },
+ {
+ "epoch": 16.083842794759825,
+ "grad_norm": 0.012737024575471878,
+ "learning_rate": 0.002380295362307836,
+ "loss": 4.612064361572266,
+ "step": 1158
+ },
+ {
+ "epoch": 16.097816593886463,
+ "grad_norm": 0.012992842122912407,
+ "learning_rate": 0.002379211747101261,
+ "loss": 4.611995697021484,
+ "step": 1159
+ },
+ {
+ "epoch": 16.111790393013102,
+ "grad_norm": 0.010210646316409111,
+ "learning_rate": 0.0023781274356148912,
+ "loss": 4.6055707931518555,
+ "step": 1160
+ },
+ {
+ "epoch": 16.125764192139737,
+ "grad_norm": 0.010265499353408813,
+ "learning_rate": 0.0023770424287223365,
+ "loss": 4.637089729309082,
+ "step": 1161
+ },
+ {
+ "epoch": 16.139737991266376,
+ "grad_norm": 0.010174009948968887,
+ "learning_rate": 0.002375956727297768,
+ "loss": 4.620571136474609,
+ "step": 1162
+ },
+ {
+ "epoch": 16.153711790393015,
+ "grad_norm": 0.009512446820735931,
+ "learning_rate": 0.0023748703322159143,
+ "loss": 4.644092559814453,
+ "step": 1163
+ },
+ {
+ "epoch": 16.16768558951965,
+ "grad_norm": 0.009849252179265022,
+ "learning_rate": 0.0023737832443520655,
+ "loss": 4.625055313110352,
+ "step": 1164
+ },
+ {
+ "epoch": 16.18165938864629,
+ "grad_norm": 0.010493188165128231,
+ "learning_rate": 0.0023726954645820674,
+ "loss": 4.644410610198975,
+ "step": 1165
+ },
+ {
+ "epoch": 16.195633187772927,
+ "grad_norm": 0.022181199863553047,
+ "learning_rate": 0.002371606993782326,
+ "loss": 4.593084335327148,
+ "step": 1166
+ },
+ {
+ "epoch": 16.209606986899562,
+ "grad_norm": 0.011906919069588184,
+ "learning_rate": 0.0023705178328298016,
+ "loss": 4.6452836990356445,
+ "step": 1167
+ },
+ {
+ "epoch": 16.2235807860262,
+ "grad_norm": 0.011874495074152946,
+ "learning_rate": 0.0023694279826020122,
+ "loss": 4.599322319030762,
+ "step": 1168
+ },
+ {
+ "epoch": 16.23755458515284,
+ "grad_norm": 0.00989708211272955,
+ "learning_rate": 0.0023683374439770295,
+ "loss": 4.607342720031738,
+ "step": 1169
+ },
+ {
+ "epoch": 16.251528384279474,
+ "grad_norm": 0.009256849065423012,
+ "learning_rate": 0.0023672462178334817,
+ "loss": 4.621583461761475,
+ "step": 1170
+ },
+ {
+ "epoch": 16.265502183406113,
+ "grad_norm": 0.013187132775783539,
+ "learning_rate": 0.0023661543050505496,
+ "loss": 4.614629745483398,
+ "step": 1171
+ },
+ {
+ "epoch": 16.27947598253275,
+ "grad_norm": 0.013505950570106506,
+ "learning_rate": 0.002365061706507968,
+ "loss": 4.629693984985352,
+ "step": 1172
+ },
+ {
+ "epoch": 16.29344978165939,
+ "grad_norm": 0.012032733298838139,
+ "learning_rate": 0.0023639684230860225,
+ "loss": 4.671209335327148,
+ "step": 1173
+ },
+ {
+ "epoch": 16.307423580786025,
+ "grad_norm": 0.01278337650001049,
+ "learning_rate": 0.002362874455665553,
+ "loss": 4.648874759674072,
+ "step": 1174
+ },
+ {
+ "epoch": 16.321397379912664,
+ "grad_norm": 0.01123690977692604,
+ "learning_rate": 0.0023617798051279496,
+ "loss": 4.611470699310303,
+ "step": 1175
+ },
+ {
+ "epoch": 16.335371179039303,
+ "grad_norm": 0.012564311735332012,
+ "learning_rate": 0.002360684472355151,
+ "loss": 4.659026145935059,
+ "step": 1176
+ },
+ {
+ "epoch": 16.349344978165938,
+ "grad_norm": 0.01206638477742672,
+ "learning_rate": 0.002359588458229649,
+ "loss": 4.642566204071045,
+ "step": 1177
+ },
+ {
+ "epoch": 16.363318777292577,
+ "grad_norm": 0.010300020687282085,
+ "learning_rate": 0.0023584917636344806,
+ "loss": 4.629585266113281,
+ "step": 1178
+ },
+ {
+ "epoch": 16.377292576419215,
+ "grad_norm": 0.0102265365421772,
+ "learning_rate": 0.0023573943894532336,
+ "loss": 4.609250068664551,
+ "step": 1179
+ },
+ {
+ "epoch": 16.39126637554585,
+ "grad_norm": 0.012440572492778301,
+ "learning_rate": 0.002356296336570042,
+ "loss": 4.636714458465576,
+ "step": 1180
+ },
+ {
+ "epoch": 16.40524017467249,
+ "grad_norm": 0.011830241419374943,
+ "learning_rate": 0.0023551976058695887,
+ "loss": 4.59100866317749,
+ "step": 1181
+ },
+ {
+ "epoch": 16.419213973799128,
+ "grad_norm": 0.011482699774205685,
+ "learning_rate": 0.0023540981982370997,
+ "loss": 4.620088577270508,
+ "step": 1182
+ },
+ {
+ "epoch": 16.433187772925763,
+ "grad_norm": 0.012950294651091099,
+ "learning_rate": 0.0023529981145583483,
+ "loss": 4.578657627105713,
+ "step": 1183
+ },
+ {
+ "epoch": 16.4471615720524,
+ "grad_norm": 0.012197710573673248,
+ "learning_rate": 0.002351897355719652,
+ "loss": 4.620813846588135,
+ "step": 1184
+ },
+ {
+ "epoch": 16.46113537117904,
+ "grad_norm": 0.014839230105280876,
+ "learning_rate": 0.002350795922607872,
+ "loss": 4.615731239318848,
+ "step": 1185
+ },
+ {
+ "epoch": 16.475109170305675,
+ "grad_norm": 0.016670791432261467,
+ "learning_rate": 0.002349693816110414,
+ "loss": 4.655855178833008,
+ "step": 1186
+ },
+ {
+ "epoch": 16.489082969432314,
+ "grad_norm": 0.014426100067794323,
+ "learning_rate": 0.0023485910371152247,
+ "loss": 4.660849571228027,
+ "step": 1187
+ },
+ {
+ "epoch": 16.503056768558952,
+ "grad_norm": 0.014040560461580753,
+ "learning_rate": 0.0023474875865107936,
+ "loss": 4.634917736053467,
+ "step": 1188
+ },
+ {
+ "epoch": 16.51703056768559,
+ "grad_norm": 0.018269391730427742,
+ "learning_rate": 0.0023463834651861496,
+ "loss": 4.611809253692627,
+ "step": 1189
+ },
+ {
+ "epoch": 16.531004366812226,
+ "grad_norm": 0.013855630531907082,
+ "learning_rate": 0.0023452786740308647,
+ "loss": 4.615225791931152,
+ "step": 1190
+ },
+ {
+ "epoch": 16.544978165938865,
+ "grad_norm": 0.011505477130413055,
+ "learning_rate": 0.0023441732139350487,
+ "loss": 4.62592887878418,
+ "step": 1191
+ },
+ {
+ "epoch": 16.558951965065503,
+ "grad_norm": 0.010166114196181297,
+ "learning_rate": 0.002343067085789351,
+ "loss": 4.6219024658203125,
+ "step": 1192
+ },
+ {
+ "epoch": 16.57292576419214,
+ "grad_norm": 0.009253021329641342,
+ "learning_rate": 0.0023419602904849596,
+ "loss": 4.658320426940918,
+ "step": 1193
+ },
+ {
+ "epoch": 16.586899563318777,
+ "grad_norm": 0.01027526892721653,
+ "learning_rate": 0.002340852828913598,
+ "loss": 4.589777946472168,
+ "step": 1194
+ },
+ {
+ "epoch": 16.600873362445416,
+ "grad_norm": 0.008855022490024567,
+ "learning_rate": 0.002339744701967529,
+ "loss": 4.588204383850098,
+ "step": 1195
+ },
+ {
+ "epoch": 16.61484716157205,
+ "grad_norm": 0.00965463649481535,
+ "learning_rate": 0.002338635910539551,
+ "loss": 4.648177146911621,
+ "step": 1196
+ },
+ {
+ "epoch": 16.62882096069869,
+ "grad_norm": 0.009073720313608646,
+ "learning_rate": 0.0023375264555229956,
+ "loss": 4.585582733154297,
+ "step": 1197
+ },
+ {
+ "epoch": 16.64279475982533,
+ "grad_norm": 0.010105383582413197,
+ "learning_rate": 0.0023364163378117317,
+ "loss": 4.612174987792969,
+ "step": 1198
+ },
+ {
+ "epoch": 16.656768558951963,
+ "grad_norm": 0.015307556837797165,
+ "learning_rate": 0.002335305558300161,
+ "loss": 4.591984748840332,
+ "step": 1199
+ },
+ {
+ "epoch": 16.670742358078602,
+ "grad_norm": 0.016716254875063896,
+ "learning_rate": 0.002334194117883219,
+ "loss": 4.659285545349121,
+ "step": 1200
+ },
+ {
+ "epoch": 16.68471615720524,
+ "grad_norm": 0.013281598687171936,
+ "learning_rate": 0.0023330820174563725,
+ "loss": 4.645829200744629,
+ "step": 1201
+ },
+ {
+ "epoch": 16.69868995633188,
+ "grad_norm": 0.014416513964533806,
+ "learning_rate": 0.0023319692579156204,
+ "loss": 4.584893226623535,
+ "step": 1202
+ },
+ {
+ "epoch": 16.712663755458514,
+ "grad_norm": 0.016326071694493294,
+ "learning_rate": 0.002330855840157494,
+ "loss": 4.633223533630371,
+ "step": 1203
+ },
+ {
+ "epoch": 16.726637554585153,
+ "grad_norm": 0.012851824052631855,
+ "learning_rate": 0.002329741765079053,
+ "loss": 4.580986499786377,
+ "step": 1204
+ },
+ {
+ "epoch": 16.74061135371179,
+ "grad_norm": 0.015444506891071796,
+ "learning_rate": 0.0023286270335778876,
+ "loss": 4.615891456604004,
+ "step": 1205
+ },
+ {
+ "epoch": 16.754585152838427,
+ "grad_norm": 0.01391979493200779,
+ "learning_rate": 0.0023275116465521174,
+ "loss": 4.628403186798096,
+ "step": 1206
+ },
+ {
+ "epoch": 16.768558951965066,
+ "grad_norm": 0.012138686142861843,
+ "learning_rate": 0.0023263956049003895,
+ "loss": 4.6259942054748535,
+ "step": 1207
+ },
+ {
+ "epoch": 16.782532751091704,
+ "grad_norm": 0.013741672039031982,
+ "learning_rate": 0.0023252789095218785,
+ "loss": 4.6056742668151855,
+ "step": 1208
+ },
+ {
+ "epoch": 16.79650655021834,
+ "grad_norm": 0.01103319600224495,
+ "learning_rate": 0.002324161561316285,
+ "loss": 4.621670246124268,
+ "step": 1209
+ },
+ {
+ "epoch": 16.810480349344978,
+ "grad_norm": 0.010333316400647163,
+ "learning_rate": 0.0023230435611838367,
+ "loss": 4.620484352111816,
+ "step": 1210
+ },
+ {
+ "epoch": 16.824454148471617,
+ "grad_norm": 0.011063606478273869,
+ "learning_rate": 0.0023219249100252864,
+ "loss": 4.64952278137207,
+ "step": 1211
+ },
+ {
+ "epoch": 16.83842794759825,
+ "grad_norm": 0.009984857402741909,
+ "learning_rate": 0.0023208056087419106,
+ "loss": 4.572571754455566,
+ "step": 1212
+ },
+ {
+ "epoch": 16.85240174672489,
+ "grad_norm": 0.009986856952309608,
+ "learning_rate": 0.002319685658235511,
+ "loss": 4.570366382598877,
+ "step": 1213
+ },
+ {
+ "epoch": 16.86637554585153,
+ "grad_norm": 0.010905452072620392,
+ "learning_rate": 0.0023185650594084103,
+ "loss": 4.603614807128906,
+ "step": 1214
+ },
+ {
+ "epoch": 16.880349344978168,
+ "grad_norm": 0.012800061143934727,
+ "learning_rate": 0.002317443813163456,
+ "loss": 4.562705039978027,
+ "step": 1215
+ },
+ {
+ "epoch": 16.894323144104803,
+ "grad_norm": 0.01252608560025692,
+ "learning_rate": 0.002316321920404015,
+ "loss": 4.599016189575195,
+ "step": 1216
+ },
+ {
+ "epoch": 16.90829694323144,
+ "grad_norm": 0.011305592954158783,
+ "learning_rate": 0.002315199382033976,
+ "loss": 4.644617557525635,
+ "step": 1217
+ },
+ {
+ "epoch": 16.92227074235808,
+ "grad_norm": 0.010594413615763187,
+ "learning_rate": 0.002314076198957749,
+ "loss": 4.569808006286621,
+ "step": 1218
+ },
+ {
+ "epoch": 16.936244541484715,
+ "grad_norm": 0.009859138168394566,
+ "learning_rate": 0.002312952372080261,
+ "loss": 4.5890398025512695,
+ "step": 1219
+ },
+ {
+ "epoch": 16.950218340611354,
+ "grad_norm": 0.010982917621731758,
+ "learning_rate": 0.00231182790230696,
+ "loss": 4.606348991394043,
+ "step": 1220
+ },
+ {
+ "epoch": 16.964192139737992,
+ "grad_norm": 0.010252755135297775,
+ "learning_rate": 0.00231070279054381,
+ "loss": 4.58942174911499,
+ "step": 1221
+ },
+ {
+ "epoch": 16.978165938864628,
+ "grad_norm": 0.01110257301479578,
+ "learning_rate": 0.0023095770376972947,
+ "loss": 4.599823474884033,
+ "step": 1222
+ },
+ {
+ "epoch": 16.992139737991266,
+ "grad_norm": 0.01003815233707428,
+ "learning_rate": 0.0023084506446744103,
+ "loss": 4.569343090057373,
+ "step": 1223
+ },
+ {
+ "epoch": 17.0,
+ "grad_norm": 0.010517258197069168,
+ "learning_rate": 0.0023073236123826745,
+ "loss": 4.646301746368408,
+ "step": 1224
+ },
+ {
+ "epoch": 17.0,
+ "eval_loss": 4.804611682891846,
+ "eval_runtime": 51.6279,
+ "eval_samples_per_second": 47.3,
+ "eval_steps_per_second": 1.491,
+ "step": 1224
+ },
+ {
+ "epoch": 17.01397379912664,
+ "grad_norm": 0.011439066380262375,
+ "learning_rate": 0.0023061959417301145,
+ "loss": 4.563335418701172,
+ "step": 1225
+ },
+ {
+ "epoch": 17.027947598253274,
+ "grad_norm": 0.01073157787322998,
+ "learning_rate": 0.002305067633625275,
+ "loss": 4.540966987609863,
+ "step": 1226
+ },
+ {
+ "epoch": 17.041921397379912,
+ "grad_norm": 0.011391228064894676,
+ "learning_rate": 0.0023039386889772136,
+ "loss": 4.555583953857422,
+ "step": 1227
+ },
+ {
+ "epoch": 17.05589519650655,
+ "grad_norm": 0.012814595364034176,
+ "learning_rate": 0.0023028091086955,
+ "loss": 4.517425537109375,
+ "step": 1228
+ },
+ {
+ "epoch": 17.069868995633186,
+ "grad_norm": 0.012994988821446896,
+ "learning_rate": 0.002301678893690218,
+ "loss": 4.538477897644043,
+ "step": 1229
+ },
+ {
+ "epoch": 17.083842794759825,
+ "grad_norm": 0.01460991520434618,
+ "learning_rate": 0.00230054804487196,
+ "loss": 4.571145057678223,
+ "step": 1230
+ },
+ {
+ "epoch": 17.097816593886463,
+ "grad_norm": 0.013036075979471207,
+ "learning_rate": 0.0022994165631518306,
+ "loss": 4.519233703613281,
+ "step": 1231
+ },
+ {
+ "epoch": 17.111790393013102,
+ "grad_norm": 0.011917057447135448,
+ "learning_rate": 0.0022982844494414444,
+ "loss": 4.5071587562561035,
+ "step": 1232
+ },
+ {
+ "epoch": 17.125764192139737,
+ "grad_norm": 0.013928866013884544,
+ "learning_rate": 0.0022971517046529257,
+ "loss": 4.531422138214111,
+ "step": 1233
+ },
+ {
+ "epoch": 17.139737991266376,
+ "grad_norm": 0.01407038513571024,
+ "learning_rate": 0.0022960183296989053,
+ "loss": 4.511205196380615,
+ "step": 1234
+ },
+ {
+ "epoch": 17.153711790393015,
+ "grad_norm": 0.019396567717194557,
+ "learning_rate": 0.0022948843254925245,
+ "loss": 4.530007362365723,
+ "step": 1235
+ },
+ {
+ "epoch": 17.16768558951965,
+ "grad_norm": 0.014425835572183132,
+ "learning_rate": 0.002293749692947428,
+ "loss": 4.530138969421387,
+ "step": 1236
+ },
+ {
+ "epoch": 17.18165938864629,
+ "grad_norm": 0.01600247249007225,
+ "learning_rate": 0.00229261443297777,
+ "loss": 4.5643391609191895,
+ "step": 1237
+ },
+ {
+ "epoch": 17.195633187772927,
+ "grad_norm": 0.018238037824630737,
+ "learning_rate": 0.0022914785464982095,
+ "loss": 4.540779113769531,
+ "step": 1238
+ },
+ {
+ "epoch": 17.209606986899562,
+ "grad_norm": 0.015427535399794579,
+ "learning_rate": 0.0022903420344239085,
+ "loss": 4.506287097930908,
+ "step": 1239
+ },
+ {
+ "epoch": 17.2235807860262,
+ "grad_norm": 0.016816625371575356,
+ "learning_rate": 0.0022892048976705347,
+ "loss": 4.537919521331787,
+ "step": 1240
+ },
+ {
+ "epoch": 17.23755458515284,
+ "grad_norm": 0.01999717950820923,
+ "learning_rate": 0.002288067137154259,
+ "loss": 4.558426380157471,
+ "step": 1241
+ },
+ {
+ "epoch": 17.251528384279474,
+ "grad_norm": 0.020884918048977852,
+ "learning_rate": 0.002286928753791753,
+ "loss": 4.592915058135986,
+ "step": 1242
+ },
+ {
+ "epoch": 17.265502183406113,
+ "grad_norm": 0.01620141603052616,
+ "learning_rate": 0.002285789748500194,
+ "loss": 4.553356170654297,
+ "step": 1243
+ },
+ {
+ "epoch": 17.27947598253275,
+ "grad_norm": 0.018495604395866394,
+ "learning_rate": 0.0022846501221972567,
+ "loss": 4.573139190673828,
+ "step": 1244
+ },
+ {
+ "epoch": 17.29344978165939,
+ "grad_norm": 0.0156320258975029,
+ "learning_rate": 0.0022835098758011174,
+ "loss": 4.585921764373779,
+ "step": 1245
+ },
+ {
+ "epoch": 17.307423580786025,
+ "grad_norm": 0.014515317976474762,
+ "learning_rate": 0.002282369010230452,
+ "loss": 4.5532002449035645,
+ "step": 1246
+ },
+ {
+ "epoch": 17.321397379912664,
+ "grad_norm": 0.014826556667685509,
+ "learning_rate": 0.0022812275264044355,
+ "loss": 4.560998916625977,
+ "step": 1247
+ },
+ {
+ "epoch": 17.335371179039303,
+ "grad_norm": 0.013240831904113293,
+ "learning_rate": 0.00228008542524274,
+ "loss": 4.531787872314453,
+ "step": 1248
+ },
+ {
+ "epoch": 17.349344978165938,
+ "grad_norm": 0.013814719393849373,
+ "learning_rate": 0.0022789427076655372,
+ "loss": 4.611872673034668,
+ "step": 1249
+ },
+ {
+ "epoch": 17.363318777292577,
+ "grad_norm": 0.012246561236679554,
+ "learning_rate": 0.0022777993745934934,
+ "loss": 4.5565385818481445,
+ "step": 1250
+ },
+ {
+ "epoch": 17.377292576419215,
+ "grad_norm": 0.01206218171864748,
+ "learning_rate": 0.002276655426947771,
+ "loss": 4.570852279663086,
+ "step": 1251
+ },
+ {
+ "epoch": 17.39126637554585,
+ "grad_norm": 0.01007329672574997,
+ "learning_rate": 0.0022755108656500275,
+ "loss": 4.553400039672852,
+ "step": 1252
+ },
+ {
+ "epoch": 17.40524017467249,
+ "grad_norm": 0.010361720807850361,
+ "learning_rate": 0.0022743656916224173,
+ "loss": 4.527509689331055,
+ "step": 1253
+ },
+ {
+ "epoch": 17.419213973799128,
+ "grad_norm": 0.009456256404519081,
+ "learning_rate": 0.0022732199057875846,
+ "loss": 4.520236015319824,
+ "step": 1254
+ },
+ {
+ "epoch": 17.433187772925763,
+ "grad_norm": 0.009310049936175346,
+ "learning_rate": 0.002272073509068669,
+ "loss": 4.518343925476074,
+ "step": 1255
+ },
+ {
+ "epoch": 17.4471615720524,
+ "grad_norm": 0.00927449855953455,
+ "learning_rate": 0.0022709265023893015,
+ "loss": 4.559690475463867,
+ "step": 1256
+ },
+ {
+ "epoch": 17.46113537117904,
+ "grad_norm": 0.008658594451844692,
+ "learning_rate": 0.002269778886673605,
+ "loss": 4.513876914978027,
+ "step": 1257
+ },
+ {
+ "epoch": 17.475109170305675,
+ "grad_norm": 0.008928220719099045,
+ "learning_rate": 0.0022686306628461918,
+ "loss": 4.545050621032715,
+ "step": 1258
+ },
+ {
+ "epoch": 17.489082969432314,
+ "grad_norm": 0.008046358823776245,
+ "learning_rate": 0.002267481831832166,
+ "loss": 4.54866361618042,
+ "step": 1259
+ },
+ {
+ "epoch": 17.503056768558952,
+ "grad_norm": 0.008318924345076084,
+ "learning_rate": 0.00226633239455712,
+ "loss": 4.546968936920166,
+ "step": 1260
+ },
+ {
+ "epoch": 17.51703056768559,
+ "grad_norm": 0.008808869868516922,
+ "learning_rate": 0.0022651823519471343,
+ "loss": 4.50877571105957,
+ "step": 1261
+ },
+ {
+ "epoch": 17.531004366812226,
+ "grad_norm": 0.008767383173108101,
+ "learning_rate": 0.0022640317049287774,
+ "loss": 4.57666540145874,
+ "step": 1262
+ },
+ {
+ "epoch": 17.544978165938865,
+ "grad_norm": 0.009351354092359543,
+ "learning_rate": 0.0022628804544291047,
+ "loss": 4.511702537536621,
+ "step": 1263
+ },
+ {
+ "epoch": 17.558951965065503,
+ "grad_norm": 0.010137057863175869,
+ "learning_rate": 0.002261728601375659,
+ "loss": 4.539676189422607,
+ "step": 1264
+ },
+ {
+ "epoch": 17.57292576419214,
+ "grad_norm": 0.00882046204060316,
+ "learning_rate": 0.002260576146696466,
+ "loss": 4.533196449279785,
+ "step": 1265
+ },
+ {
+ "epoch": 17.586899563318777,
+ "grad_norm": 0.009354782290756702,
+ "learning_rate": 0.0022594230913200388,
+ "loss": 4.515488624572754,
+ "step": 1266
+ },
+ {
+ "epoch": 17.600873362445416,
+ "grad_norm": 0.009241201914846897,
+ "learning_rate": 0.002258269436175372,
+ "loss": 4.542706489562988,
+ "step": 1267
+ },
+ {
+ "epoch": 17.61484716157205,
+ "grad_norm": 0.009252829477190971,
+ "learning_rate": 0.002257115182191947,
+ "loss": 4.592660903930664,
+ "step": 1268
+ },
+ {
+ "epoch": 17.62882096069869,
+ "grad_norm": 0.011243131943047047,
+ "learning_rate": 0.0022559603302997225,
+ "loss": 4.527804851531982,
+ "step": 1269
+ },
+ {
+ "epoch": 17.64279475982533,
+ "grad_norm": 0.011032823473215103,
+ "learning_rate": 0.0022548048814291437,
+ "loss": 4.540122032165527,
+ "step": 1270
+ },
+ {
+ "epoch": 17.656768558951963,
+ "grad_norm": 0.010495275259017944,
+ "learning_rate": 0.0022536488365111347,
+ "loss": 4.532835006713867,
+ "step": 1271
+ },
+ {
+ "epoch": 17.670742358078602,
+ "grad_norm": 0.009992090053856373,
+ "learning_rate": 0.0022524921964770994,
+ "loss": 4.521217346191406,
+ "step": 1272
+ },
+ {
+ "epoch": 17.68471615720524,
+ "grad_norm": 0.01141198631376028,
+ "learning_rate": 0.0022513349622589225,
+ "loss": 4.519709587097168,
+ "step": 1273
+ },
+ {
+ "epoch": 17.69868995633188,
+ "grad_norm": 0.013087889179587364,
+ "learning_rate": 0.002250177134788966,
+ "loss": 4.547289848327637,
+ "step": 1274
+ },
+ {
+ "epoch": 17.712663755458514,
+ "grad_norm": 0.01328477542847395,
+ "learning_rate": 0.0022490187150000706,
+ "loss": 4.527865409851074,
+ "step": 1275
+ },
+ {
+ "epoch": 17.726637554585153,
+ "grad_norm": 0.01033794041723013,
+ "learning_rate": 0.002247859703825555,
+ "loss": 4.54740047454834,
+ "step": 1276
+ },
+ {
+ "epoch": 17.74061135371179,
+ "grad_norm": 0.010522300377488136,
+ "learning_rate": 0.002246700102199213,
+ "loss": 4.538840293884277,
+ "step": 1277
+ },
+ {
+ "epoch": 17.754585152838427,
+ "grad_norm": 0.012151090428233147,
+ "learning_rate": 0.0022455399110553145,
+ "loss": 4.536865234375,
+ "step": 1278
+ },
+ {
+ "epoch": 17.768558951965066,
+ "grad_norm": 0.010553102008998394,
+ "learning_rate": 0.0022443791313286054,
+ "loss": 4.5348334312438965,
+ "step": 1279
+ },
+ {
+ "epoch": 17.782532751091704,
+ "grad_norm": 0.011433709412813187,
+ "learning_rate": 0.0022432177639543035,
+ "loss": 4.510526657104492,
+ "step": 1280
+ },
+ {
+ "epoch": 17.79650655021834,
+ "grad_norm": 0.01189691387116909,
+ "learning_rate": 0.002242055809868103,
+ "loss": 4.530807018280029,
+ "step": 1281
+ },
+ {
+ "epoch": 17.810480349344978,
+ "grad_norm": 0.010031254962086678,
+ "learning_rate": 0.0022408932700061683,
+ "loss": 4.496850967407227,
+ "step": 1282
+ },
+ {
+ "epoch": 17.824454148471617,
+ "grad_norm": 0.010129331611096859,
+ "learning_rate": 0.002239730145305138,
+ "loss": 4.551287651062012,
+ "step": 1283
+ },
+ {
+ "epoch": 17.83842794759825,
+ "grad_norm": 0.010430586524307728,
+ "learning_rate": 0.0022385664367021185,
+ "loss": 4.542672157287598,
+ "step": 1284
+ },
+ {
+ "epoch": 17.85240174672489,
+ "grad_norm": 0.011774476617574692,
+ "learning_rate": 0.002237402145134691,
+ "loss": 4.537034034729004,
+ "step": 1285
+ },
+ {
+ "epoch": 17.86637554585153,
+ "grad_norm": 0.01184659544378519,
+ "learning_rate": 0.002236237271540903,
+ "loss": 4.521193504333496,
+ "step": 1286
+ },
+ {
+ "epoch": 17.880349344978168,
+ "grad_norm": 0.011327247135341167,
+ "learning_rate": 0.0022350718168592724,
+ "loss": 4.49426794052124,
+ "step": 1287
+ },
+ {
+ "epoch": 17.894323144104803,
+ "grad_norm": 0.010767918080091476,
+ "learning_rate": 0.0022339057820287855,
+ "loss": 4.531225204467773,
+ "step": 1288
+ },
+ {
+ "epoch": 17.90829694323144,
+ "grad_norm": 0.011195424944162369,
+ "learning_rate": 0.002232739167988894,
+ "loss": 4.492671012878418,
+ "step": 1289
+ },
+ {
+ "epoch": 17.92227074235808,
+ "grad_norm": 0.010394908487796783,
+ "learning_rate": 0.0022315719756795197,
+ "loss": 4.523616790771484,
+ "step": 1290
+ },
+ {
+ "epoch": 17.936244541484715,
+ "grad_norm": 0.010277998633682728,
+ "learning_rate": 0.002230404206041047,
+ "loss": 4.487832069396973,
+ "step": 1291
+ },
+ {
+ "epoch": 17.950218340611354,
+ "grad_norm": 0.010955783538520336,
+ "learning_rate": 0.0022292358600143275,
+ "loss": 4.526865005493164,
+ "step": 1292
+ },
+ {
+ "epoch": 17.964192139737992,
+ "grad_norm": 0.010611076839268208,
+ "learning_rate": 0.002228066938540676,
+ "loss": 4.498682022094727,
+ "step": 1293
+ },
+ {
+ "epoch": 17.978165938864628,
+ "grad_norm": 0.010008828714489937,
+ "learning_rate": 0.0022268974425618724,
+ "loss": 4.524019241333008,
+ "step": 1294
+ },
+ {
+ "epoch": 17.992139737991266,
+ "grad_norm": 0.008973540738224983,
+ "learning_rate": 0.002225727373020157,
+ "loss": 4.512996673583984,
+ "step": 1295
+ },
+ {
+ "epoch": 18.0,
+ "grad_norm": 0.011754963546991348,
+ "learning_rate": 0.002224556730858235,
+ "loss": 4.496376991271973,
+ "step": 1296
+ },
+ {
+ "epoch": 18.0,
+ "eval_loss": 4.744190216064453,
+ "eval_runtime": 51.6572,
+ "eval_samples_per_second": 47.273,
+ "eval_steps_per_second": 1.491,
+ "step": 1296
+ },
+ {
+ "epoch": 18.01397379912664,
+ "grad_norm": 0.012094628997147083,
+ "learning_rate": 0.0022233855170192703,
+ "loss": 4.443447113037109,
+ "step": 1297
+ },
+ {
+ "epoch": 18.027947598253274,
+ "grad_norm": 0.014347990974783897,
+ "learning_rate": 0.00222221373244689,
+ "loss": 4.485690116882324,
+ "step": 1298
+ },
+ {
+ "epoch": 18.041921397379912,
+ "grad_norm": 0.021286984905600548,
+ "learning_rate": 0.0022210413780851786,
+ "loss": 4.487505912780762,
+ "step": 1299
+ },
+ {
+ "epoch": 18.05589519650655,
+ "grad_norm": 0.020273910835385323,
+ "learning_rate": 0.0022198684548786827,
+ "loss": 4.473119735717773,
+ "step": 1300
+ },
+ {
+ "epoch": 18.069868995633186,
+ "grad_norm": 0.017642971128225327,
+ "learning_rate": 0.002218694963772403,
+ "loss": 4.440615653991699,
+ "step": 1301
+ },
+ {
+ "epoch": 18.083842794759825,
+ "grad_norm": 0.02075147069990635,
+ "learning_rate": 0.0022175209057118017,
+ "loss": 4.527026176452637,
+ "step": 1302
+ },
+ {
+ "epoch": 18.097816593886463,
+ "grad_norm": 0.016479508951306343,
+ "learning_rate": 0.002216346281642796,
+ "loss": 4.508321762084961,
+ "step": 1303
+ },
+ {
+ "epoch": 18.111790393013102,
+ "grad_norm": 0.017978593707084656,
+ "learning_rate": 0.002215171092511759,
+ "loss": 4.500985145568848,
+ "step": 1304
+ },
+ {
+ "epoch": 18.125764192139737,
+ "grad_norm": 0.01739897020161152,
+ "learning_rate": 0.002213995339265519,
+ "loss": 4.474225044250488,
+ "step": 1305
+ },
+ {
+ "epoch": 18.139737991266376,
+ "grad_norm": 0.015556776896119118,
+ "learning_rate": 0.0022128190228513597,
+ "loss": 4.444849014282227,
+ "step": 1306
+ },
+ {
+ "epoch": 18.153711790393015,
+ "grad_norm": 0.016166578978300095,
+ "learning_rate": 0.002211642144217018,
+ "loss": 4.5026726722717285,
+ "step": 1307
+ },
+ {
+ "epoch": 18.16768558951965,
+ "grad_norm": 0.01534261740744114,
+ "learning_rate": 0.0022104647043106832,
+ "loss": 4.484527587890625,
+ "step": 1308
+ },
+ {
+ "epoch": 18.18165938864629,
+ "grad_norm": 0.013517106883227825,
+ "learning_rate": 0.002209286704080998,
+ "loss": 4.453907012939453,
+ "step": 1309
+ },
+ {
+ "epoch": 18.195633187772927,
+ "grad_norm": 0.012562787160277367,
+ "learning_rate": 0.0022081081444770556,
+ "loss": 4.471528053283691,
+ "step": 1310
+ },
+ {
+ "epoch": 18.209606986899562,
+ "grad_norm": 0.009947682730853558,
+ "learning_rate": 0.0022069290264483997,
+ "loss": 4.504153728485107,
+ "step": 1311
+ },
+ {
+ "epoch": 18.2235807860262,
+ "grad_norm": 0.01169571466743946,
+ "learning_rate": 0.002205749350945026,
+ "loss": 4.447345733642578,
+ "step": 1312
+ },
+ {
+ "epoch": 18.23755458515284,
+ "grad_norm": 0.011394352652132511,
+ "learning_rate": 0.0022045691189173755,
+ "loss": 4.447491645812988,
+ "step": 1313
+ },
+ {
+ "epoch": 18.251528384279474,
+ "grad_norm": 0.011168593540787697,
+ "learning_rate": 0.0022033883313163417,
+ "loss": 4.441639423370361,
+ "step": 1314
+ },
+ {
+ "epoch": 18.265502183406113,
+ "grad_norm": 0.009503341279923916,
+ "learning_rate": 0.002202206989093263,
+ "loss": 4.461601257324219,
+ "step": 1315
+ },
+ {
+ "epoch": 18.27947598253275,
+ "grad_norm": 0.0112199941650033,
+ "learning_rate": 0.002201025093199926,
+ "loss": 4.444095134735107,
+ "step": 1316
+ },
+ {
+ "epoch": 18.29344978165939,
+ "grad_norm": 0.010505333542823792,
+ "learning_rate": 0.002199842644588562,
+ "loss": 4.460987091064453,
+ "step": 1317
+ },
+ {
+ "epoch": 18.307423580786025,
+ "grad_norm": 0.009494813159108162,
+ "learning_rate": 0.0021986596442118496,
+ "loss": 4.460158348083496,
+ "step": 1318
+ },
+ {
+ "epoch": 18.321397379912664,
+ "grad_norm": 0.01063452661037445,
+ "learning_rate": 0.00219747609302291,
+ "loss": 4.489625453948975,
+ "step": 1319
+ },
+ {
+ "epoch": 18.335371179039303,
+ "grad_norm": 0.012194888666272163,
+ "learning_rate": 0.0021962919919753087,
+ "loss": 4.441605567932129,
+ "step": 1320
+ },
+ {
+ "epoch": 18.349344978165938,
+ "grad_norm": 0.010613647289574146,
+ "learning_rate": 0.0021951073420230557,
+ "loss": 4.475821018218994,
+ "step": 1321
+ },
+ {
+ "epoch": 18.363318777292577,
+ "grad_norm": 0.009643700905144215,
+ "learning_rate": 0.002193922144120601,
+ "loss": 4.467228889465332,
+ "step": 1322
+ },
+ {
+ "epoch": 18.377292576419215,
+ "grad_norm": 0.00912613607943058,
+ "learning_rate": 0.0021927363992228373,
+ "loss": 4.455435752868652,
+ "step": 1323
+ },
+ {
+ "epoch": 18.39126637554585,
+ "grad_norm": 0.009371735155582428,
+ "learning_rate": 0.0021915501082850988,
+ "loss": 4.479915142059326,
+ "step": 1324
+ },
+ {
+ "epoch": 18.40524017467249,
+ "grad_norm": 0.008617875166237354,
+ "learning_rate": 0.002190363272263158,
+ "loss": 4.415642738342285,
+ "step": 1325
+ },
+ {
+ "epoch": 18.419213973799128,
+ "grad_norm": 0.009571383707225323,
+ "learning_rate": 0.002189175892113227,
+ "loss": 4.439854621887207,
+ "step": 1326
+ },
+ {
+ "epoch": 18.433187772925763,
+ "grad_norm": 0.00782635249197483,
+ "learning_rate": 0.002187987968791957,
+ "loss": 4.484311103820801,
+ "step": 1327
+ },
+ {
+ "epoch": 18.4471615720524,
+ "grad_norm": 0.008408057503402233,
+ "learning_rate": 0.0021867995032564358,
+ "loss": 4.46577262878418,
+ "step": 1328
+ },
+ {
+ "epoch": 18.46113537117904,
+ "grad_norm": 0.009641910903155804,
+ "learning_rate": 0.00218561049646419,
+ "loss": 4.437408447265625,
+ "step": 1329
+ },
+ {
+ "epoch": 18.475109170305675,
+ "grad_norm": 0.008498182520270348,
+ "learning_rate": 0.0021844209493731784,
+ "loss": 4.442599296569824,
+ "step": 1330
+ },
+ {
+ "epoch": 18.489082969432314,
+ "grad_norm": 0.009289798326790333,
+ "learning_rate": 0.0021832308629418,
+ "loss": 4.517641067504883,
+ "step": 1331
+ },
+ {
+ "epoch": 18.503056768558952,
+ "grad_norm": 0.009360183961689472,
+ "learning_rate": 0.0021820402381288848,
+ "loss": 4.474143981933594,
+ "step": 1332
+ },
+ {
+ "epoch": 18.51703056768559,
+ "grad_norm": 0.008764738216996193,
+ "learning_rate": 0.0021808490758936983,
+ "loss": 4.4376702308654785,
+ "step": 1333
+ },
+ {
+ "epoch": 18.531004366812226,
+ "grad_norm": 0.009673701599240303,
+ "learning_rate": 0.0021796573771959375,
+ "loss": 4.435220718383789,
+ "step": 1334
+ },
+ {
+ "epoch": 18.544978165938865,
+ "grad_norm": 0.010852573439478874,
+ "learning_rate": 0.0021784651429957334,
+ "loss": 4.465187072753906,
+ "step": 1335
+ },
+ {
+ "epoch": 18.558951965065503,
+ "grad_norm": 0.011292314156889915,
+ "learning_rate": 0.0021772723742536464,
+ "loss": 4.458176136016846,
+ "step": 1336
+ },
+ {
+ "epoch": 18.57292576419214,
+ "grad_norm": 0.01071334257721901,
+ "learning_rate": 0.0021760790719306702,
+ "loss": 4.425498962402344,
+ "step": 1337
+ },
+ {
+ "epoch": 18.586899563318777,
+ "grad_norm": 0.01505239401012659,
+ "learning_rate": 0.0021748852369882265,
+ "loss": 4.463129043579102,
+ "step": 1338
+ },
+ {
+ "epoch": 18.600873362445416,
+ "grad_norm": 0.011774689890444279,
+ "learning_rate": 0.0021736908703881663,
+ "loss": 4.462577819824219,
+ "step": 1339
+ },
+ {
+ "epoch": 18.61484716157205,
+ "grad_norm": 0.011241118423640728,
+ "learning_rate": 0.002172495973092769,
+ "loss": 4.452583312988281,
+ "step": 1340
+ },
+ {
+ "epoch": 18.62882096069869,
+ "grad_norm": 0.01072910986840725,
+ "learning_rate": 0.002171300546064742,
+ "loss": 4.466248035430908,
+ "step": 1341
+ },
+ {
+ "epoch": 18.64279475982533,
+ "grad_norm": 0.01021308358758688,
+ "learning_rate": 0.002170104590267219,
+ "loss": 4.485857963562012,
+ "step": 1342
+ },
+ {
+ "epoch": 18.656768558951963,
+ "grad_norm": 0.012630444951355457,
+ "learning_rate": 0.00216890810666376,
+ "loss": 4.442112922668457,
+ "step": 1343
+ },
+ {
+ "epoch": 18.670742358078602,
+ "grad_norm": 0.013602569699287415,
+ "learning_rate": 0.0021677110962183515,
+ "loss": 4.45171594619751,
+ "step": 1344
+ },
+ {
+ "epoch": 18.68471615720524,
+ "grad_norm": 0.013808641582727432,
+ "learning_rate": 0.0021665135598954014,
+ "loss": 4.468818187713623,
+ "step": 1345
+ },
+ {
+ "epoch": 18.69868995633188,
+ "grad_norm": 0.011510699987411499,
+ "learning_rate": 0.002165315498659744,
+ "loss": 4.439584255218506,
+ "step": 1346
+ },
+ {
+ "epoch": 18.712663755458514,
+ "grad_norm": 0.010502424091100693,
+ "learning_rate": 0.0021641169134766353,
+ "loss": 4.478377342224121,
+ "step": 1347
+ },
+ {
+ "epoch": 18.726637554585153,
+ "grad_norm": 0.009953407570719719,
+ "learning_rate": 0.0021629178053117542,
+ "loss": 4.474404335021973,
+ "step": 1348
+ },
+ {
+ "epoch": 18.74061135371179,
+ "grad_norm": 0.011158930137753487,
+ "learning_rate": 0.0021617181751312,
+ "loss": 4.4549560546875,
+ "step": 1349
+ },
+ {
+ "epoch": 18.754585152838427,
+ "grad_norm": 0.010431940667331219,
+ "learning_rate": 0.002160518023901493,
+ "loss": 4.444911003112793,
+ "step": 1350
+ },
+ {
+ "epoch": 18.768558951965066,
+ "grad_norm": 0.009100335650146008,
+ "learning_rate": 0.0021593173525895734,
+ "loss": 4.428071975708008,
+ "step": 1351
+ },
+ {
+ "epoch": 18.782532751091704,
+ "grad_norm": 0.009362847544252872,
+ "learning_rate": 0.0021581161621628013,
+ "loss": 4.462827682495117,
+ "step": 1352
+ },
+ {
+ "epoch": 18.79650655021834,
+ "grad_norm": 0.009893129579722881,
+ "learning_rate": 0.0021569144535889535,
+ "loss": 4.426912307739258,
+ "step": 1353
+ },
+ {
+ "epoch": 18.810480349344978,
+ "grad_norm": 0.010524468496441841,
+ "learning_rate": 0.0021557122278362244,
+ "loss": 4.435606002807617,
+ "step": 1354
+ },
+ {
+ "epoch": 18.824454148471617,
+ "grad_norm": 0.010621445253491402,
+ "learning_rate": 0.002154509485873226,
+ "loss": 4.454901695251465,
+ "step": 1355
+ },
+ {
+ "epoch": 18.83842794759825,
+ "grad_norm": 0.010836896486580372,
+ "learning_rate": 0.0021533062286689855,
+ "loss": 4.4441728591918945,
+ "step": 1356
+ },
+ {
+ "epoch": 18.85240174672489,
+ "grad_norm": 0.011403030715882778,
+ "learning_rate": 0.0021521024571929457,
+ "loss": 4.483884811401367,
+ "step": 1357
+ },
+ {
+ "epoch": 18.86637554585153,
+ "grad_norm": 0.012522426433861256,
+ "learning_rate": 0.0021508981724149643,
+ "loss": 4.464006423950195,
+ "step": 1358
+ },
+ {
+ "epoch": 18.880349344978168,
+ "grad_norm": 0.014799179509282112,
+ "learning_rate": 0.0021496933753053105,
+ "loss": 4.460402011871338,
+ "step": 1359
+ },
+ {
+ "epoch": 18.894323144104803,
+ "grad_norm": 0.012453878298401833,
+ "learning_rate": 0.0021484880668346685,
+ "loss": 4.456899642944336,
+ "step": 1360
+ },
+ {
+ "epoch": 18.90829694323144,
+ "grad_norm": 0.012270379811525345,
+ "learning_rate": 0.002147282247974133,
+ "loss": 4.477325439453125,
+ "step": 1361
+ },
+ {
+ "epoch": 18.92227074235808,
+ "grad_norm": 0.01094560045748949,
+ "learning_rate": 0.0021460759196952105,
+ "loss": 4.422362327575684,
+ "step": 1362
+ },
+ {
+ "epoch": 18.936244541484715,
+ "grad_norm": 0.009995176456868649,
+ "learning_rate": 0.0021448690829698183,
+ "loss": 4.405127048492432,
+ "step": 1363
+ },
+ {
+ "epoch": 18.950218340611354,
+ "grad_norm": 0.014935536310076714,
+ "learning_rate": 0.002143661738770283,
+ "loss": 4.445840358734131,
+ "step": 1364
+ },
+ {
+ "epoch": 18.964192139737992,
+ "grad_norm": 0.011492030695080757,
+ "learning_rate": 0.0021424538880693393,
+ "loss": 4.459555625915527,
+ "step": 1365
+ },
+ {
+ "epoch": 18.978165938864628,
+ "grad_norm": 0.008793608285486698,
+ "learning_rate": 0.002141245531840132,
+ "loss": 4.436631202697754,
+ "step": 1366
+ },
+ {
+ "epoch": 18.992139737991266,
+ "grad_norm": 0.008894025348126888,
+ "learning_rate": 0.0021400366710562103,
+ "loss": 4.473924160003662,
+ "step": 1367
+ },
+ {
+ "epoch": 19.0,
+ "grad_norm": 0.010324863716959953,
+ "learning_rate": 0.0021388273066915327,
+ "loss": 4.475680351257324,
+ "step": 1368
+ },
+ {
+ "epoch": 19.0,
+ "eval_loss": 4.688324451446533,
+ "eval_runtime": 54.4515,
+ "eval_samples_per_second": 44.847,
+ "eval_steps_per_second": 1.414,
+ "step": 1368
+ },
+ {
+ "epoch": 19.01397379912664,
+ "grad_norm": 0.009235628880560398,
+ "learning_rate": 0.0021376174397204615,
+ "loss": 4.370064735412598,
+ "step": 1369
+ },
+ {
+ "epoch": 19.027947598253274,
+ "grad_norm": 0.008818200789391994,
+ "learning_rate": 0.0021364070711177643,
+ "loss": 4.371678352355957,
+ "step": 1370
+ },
+ {
+ "epoch": 19.041921397379912,
+ "grad_norm": 0.010333757847547531,
+ "learning_rate": 0.0021351962018586147,
+ "loss": 4.3896379470825195,
+ "step": 1371
+ },
+ {
+ "epoch": 19.05589519650655,
+ "grad_norm": 0.011938809417188168,
+ "learning_rate": 0.002133984832918587,
+ "loss": 4.361343860626221,
+ "step": 1372
+ },
+ {
+ "epoch": 19.069868995633186,
+ "grad_norm": 0.010634614154696465,
+ "learning_rate": 0.00213277296527366,
+ "loss": 4.362449645996094,
+ "step": 1373
+ },
+ {
+ "epoch": 19.083842794759825,
+ "grad_norm": 0.008545687422156334,
+ "learning_rate": 0.0021315605999002126,
+ "loss": 4.3685503005981445,
+ "step": 1374
+ },
+ {
+ "epoch": 19.097816593886463,
+ "grad_norm": 0.009159795008599758,
+ "learning_rate": 0.002130347737775027,
+ "loss": 4.379969596862793,
+ "step": 1375
+ },
+ {
+ "epoch": 19.111790393013102,
+ "grad_norm": 0.010464153252542019,
+ "learning_rate": 0.0021291343798752835,
+ "loss": 4.376113414764404,
+ "step": 1376
+ },
+ {
+ "epoch": 19.125764192139737,
+ "grad_norm": 0.011010527610778809,
+ "learning_rate": 0.0021279205271785622,
+ "loss": 4.390711784362793,
+ "step": 1377
+ },
+ {
+ "epoch": 19.139737991266376,
+ "grad_norm": 0.010350867174565792,
+ "learning_rate": 0.002126706180662843,
+ "loss": 4.399683952331543,
+ "step": 1378
+ },
+ {
+ "epoch": 19.153711790393015,
+ "grad_norm": 0.009708237834274769,
+ "learning_rate": 0.002125491341306503,
+ "loss": 4.363435745239258,
+ "step": 1379
+ },
+ {
+ "epoch": 19.16768558951965,
+ "grad_norm": 0.010087115690112114,
+ "learning_rate": 0.002124276010088316,
+ "loss": 4.366623878479004,
+ "step": 1380
+ },
+ {
+ "epoch": 19.18165938864629,
+ "grad_norm": 0.009712349623441696,
+ "learning_rate": 0.002123060187987453,
+ "loss": 4.360814571380615,
+ "step": 1381
+ },
+ {
+ "epoch": 19.195633187772927,
+ "grad_norm": 0.010174986906349659,
+ "learning_rate": 0.0021218438759834785,
+ "loss": 4.347475051879883,
+ "step": 1382
+ },
+ {
+ "epoch": 19.209606986899562,
+ "grad_norm": 0.010195785202085972,
+ "learning_rate": 0.0021206270750563543,
+ "loss": 4.399251461029053,
+ "step": 1383
+ },
+ {
+ "epoch": 19.2235807860262,
+ "grad_norm": 0.010786959901452065,
+ "learning_rate": 0.0021194097861864354,
+ "loss": 4.397933006286621,
+ "step": 1384
+ },
+ {
+ "epoch": 19.23755458515284,
+ "grad_norm": 0.011476189829409122,
+ "learning_rate": 0.002118192010354468,
+ "loss": 4.3748369216918945,
+ "step": 1385
+ },
+ {
+ "epoch": 19.251528384279474,
+ "grad_norm": 0.011571013368666172,
+ "learning_rate": 0.0021169737485415927,
+ "loss": 4.388349533081055,
+ "step": 1386
+ },
+ {
+ "epoch": 19.265502183406113,
+ "grad_norm": 0.011140935122966766,
+ "learning_rate": 0.002115755001729341,
+ "loss": 4.398018836975098,
+ "step": 1387
+ },
+ {
+ "epoch": 19.27947598253275,
+ "grad_norm": 0.013216734863817692,
+ "learning_rate": 0.0021145357708996365,
+ "loss": 4.406947135925293,
+ "step": 1388
+ },
+ {
+ "epoch": 19.29344978165939,
+ "grad_norm": 0.013385126367211342,
+ "learning_rate": 0.0021133160570347893,
+ "loss": 4.41457462310791,
+ "step": 1389
+ },
+ {
+ "epoch": 19.307423580786025,
+ "grad_norm": 0.011784231290221214,
+ "learning_rate": 0.002112095861117502,
+ "loss": 4.403835296630859,
+ "step": 1390
+ },
+ {
+ "epoch": 19.321397379912664,
+ "grad_norm": 0.009813498705625534,
+ "learning_rate": 0.0021108751841308645,
+ "loss": 4.3865647315979,
+ "step": 1391
+ },
+ {
+ "epoch": 19.335371179039303,
+ "grad_norm": 0.010297936387360096,
+ "learning_rate": 0.0021096540270583533,
+ "loss": 4.395583152770996,
+ "step": 1392
+ },
+ {
+ "epoch": 19.349344978165938,
+ "grad_norm": 0.011965147219598293,
+ "learning_rate": 0.002108432390883834,
+ "loss": 4.385443687438965,
+ "step": 1393
+ },
+ {
+ "epoch": 19.363318777292577,
+ "grad_norm": 0.012212193571031094,
+ "learning_rate": 0.0021072102765915556,
+ "loss": 4.422321319580078,
+ "step": 1394
+ },
+ {
+ "epoch": 19.377292576419215,
+ "grad_norm": 0.009840975515544415,
+ "learning_rate": 0.0021059876851661544,
+ "loss": 4.3568902015686035,
+ "step": 1395
+ },
+ {
+ "epoch": 19.39126637554585,
+ "grad_norm": 0.013039806857705116,
+ "learning_rate": 0.002104764617592649,
+ "loss": 4.314867973327637,
+ "step": 1396
+ },
+ {
+ "epoch": 19.40524017467249,
+ "grad_norm": 0.015091902576386929,
+ "learning_rate": 0.0021035410748564444,
+ "loss": 4.392338752746582,
+ "step": 1397
+ },
+ {
+ "epoch": 19.419213973799128,
+ "grad_norm": 0.015448705293238163,
+ "learning_rate": 0.0021023170579433257,
+ "loss": 4.379609107971191,
+ "step": 1398
+ },
+ {
+ "epoch": 19.433187772925763,
+ "grad_norm": 0.018393104895949364,
+ "learning_rate": 0.0021010925678394622,
+ "loss": 4.418291091918945,
+ "step": 1399
+ },
+ {
+ "epoch": 19.4471615720524,
+ "grad_norm": 0.01866372860968113,
+ "learning_rate": 0.002099867605531402,
+ "loss": 4.4006452560424805,
+ "step": 1400
+ },
+ {
+ "epoch": 19.46113537117904,
+ "grad_norm": 0.01819106750190258,
+ "learning_rate": 0.002098642172006077,
+ "loss": 4.433975696563721,
+ "step": 1401
+ },
+ {
+ "epoch": 19.475109170305675,
+ "grad_norm": 0.019774602726101875,
+ "learning_rate": 0.0020974162682507952,
+ "loss": 4.417821884155273,
+ "step": 1402
+ },
+ {
+ "epoch": 19.489082969432314,
+ "grad_norm": 0.016885802149772644,
+ "learning_rate": 0.0020961898952532457,
+ "loss": 4.397073268890381,
+ "step": 1403
+ },
+ {
+ "epoch": 19.503056768558952,
+ "grad_norm": 0.015750881284475327,
+ "learning_rate": 0.0020949630540014956,
+ "loss": 4.4356608390808105,
+ "step": 1404
+ },
+ {
+ "epoch": 19.51703056768559,
+ "grad_norm": 0.015696514397859573,
+ "learning_rate": 0.002093735745483988,
+ "loss": 4.422449588775635,
+ "step": 1405
+ },
+ {
+ "epoch": 19.531004366812226,
+ "grad_norm": 0.017195021733641624,
+ "learning_rate": 0.0020925079706895435,
+ "loss": 4.443209648132324,
+ "step": 1406
+ },
+ {
+ "epoch": 19.544978165938865,
+ "grad_norm": 0.015864048153162003,
+ "learning_rate": 0.0020912797306073577,
+ "loss": 4.4340691566467285,
+ "step": 1407
+ },
+ {
+ "epoch": 19.558951965065503,
+ "grad_norm": 0.009857800789177418,
+ "learning_rate": 0.0020900510262270022,
+ "loss": 4.37935209274292,
+ "step": 1408
+ },
+ {
+ "epoch": 19.57292576419214,
+ "grad_norm": 0.013739476911723614,
+ "learning_rate": 0.0020888218585384216,
+ "loss": 4.412148475646973,
+ "step": 1409
+ },
+ {
+ "epoch": 19.586899563318777,
+ "grad_norm": 0.011667690239846706,
+ "learning_rate": 0.0020875922285319344,
+ "loss": 4.391700744628906,
+ "step": 1410
+ },
+ {
+ "epoch": 19.600873362445416,
+ "grad_norm": 0.012323460541665554,
+ "learning_rate": 0.0020863621371982302,
+ "loss": 4.423880577087402,
+ "step": 1411
+ },
+ {
+ "epoch": 19.61484716157205,
+ "grad_norm": 0.010353626683354378,
+ "learning_rate": 0.0020851315855283724,
+ "loss": 4.427382469177246,
+ "step": 1412
+ },
+ {
+ "epoch": 19.62882096069869,
+ "grad_norm": 0.0101751908659935,
+ "learning_rate": 0.002083900574513794,
+ "loss": 4.415317535400391,
+ "step": 1413
+ },
+ {
+ "epoch": 19.64279475982533,
+ "grad_norm": 0.01126549206674099,
+ "learning_rate": 0.002082669105146298,
+ "loss": 4.402498245239258,
+ "step": 1414
+ },
+ {
+ "epoch": 19.656768558951963,
+ "grad_norm": 0.011096738278865814,
+ "learning_rate": 0.0020814371784180575,
+ "loss": 4.433567047119141,
+ "step": 1415
+ },
+ {
+ "epoch": 19.670742358078602,
+ "grad_norm": 0.010944445617496967,
+ "learning_rate": 0.0020802047953216137,
+ "loss": 4.406990051269531,
+ "step": 1416
+ },
+ {
+ "epoch": 19.68471615720524,
+ "grad_norm": 0.008965318091213703,
+ "learning_rate": 0.0020789719568498747,
+ "loss": 4.3721771240234375,
+ "step": 1417
+ },
+ {
+ "epoch": 19.69868995633188,
+ "grad_norm": 0.00828113779425621,
+ "learning_rate": 0.002077738663996117,
+ "loss": 4.389828205108643,
+ "step": 1418
+ },
+ {
+ "epoch": 19.712663755458514,
+ "grad_norm": 0.008599454537034035,
+ "learning_rate": 0.002076504917753982,
+ "loss": 4.411396026611328,
+ "step": 1419
+ },
+ {
+ "epoch": 19.726637554585153,
+ "grad_norm": 0.007713889703154564,
+ "learning_rate": 0.0020752707191174762,
+ "loss": 4.355343341827393,
+ "step": 1420
+ },
+ {
+ "epoch": 19.74061135371179,
+ "grad_norm": 0.00912355538457632,
+ "learning_rate": 0.0020740360690809723,
+ "loss": 4.423413276672363,
+ "step": 1421
+ },
+ {
+ "epoch": 19.754585152838427,
+ "grad_norm": 0.008269775658845901,
+ "learning_rate": 0.0020728009686392045,
+ "loss": 4.390920639038086,
+ "step": 1422
+ },
+ {
+ "epoch": 19.768558951965066,
+ "grad_norm": 0.007461601868271828,
+ "learning_rate": 0.002071565418787272,
+ "loss": 4.365501403808594,
+ "step": 1423
+ },
+ {
+ "epoch": 19.782532751091704,
+ "grad_norm": 0.007951333187520504,
+ "learning_rate": 0.0020703294205206336,
+ "loss": 4.378884315490723,
+ "step": 1424
+ },
+ {
+ "epoch": 19.79650655021834,
+ "grad_norm": 0.007816503755748272,
+ "learning_rate": 0.002069092974835112,
+ "loss": 4.4241743087768555,
+ "step": 1425
+ },
+ {
+ "epoch": 19.810480349344978,
+ "grad_norm": 0.00833756010979414,
+ "learning_rate": 0.0020678560827268886,
+ "loss": 4.391102313995361,
+ "step": 1426
+ },
+ {
+ "epoch": 19.824454148471617,
+ "grad_norm": 0.008241684176027775,
+ "learning_rate": 0.0020666187451925055,
+ "loss": 4.374970436096191,
+ "step": 1427
+ },
+ {
+ "epoch": 19.83842794759825,
+ "grad_norm": 0.008245996199548244,
+ "learning_rate": 0.002065380963228862,
+ "loss": 4.380222320556641,
+ "step": 1428
+ },
+ {
+ "epoch": 19.85240174672489,
+ "grad_norm": 0.007378706708550453,
+ "learning_rate": 0.0020641427378332183,
+ "loss": 4.391451358795166,
+ "step": 1429
+ },
+ {
+ "epoch": 19.86637554585153,
+ "grad_norm": 0.007594958413392305,
+ "learning_rate": 0.002062904070003189,
+ "loss": 4.372774124145508,
+ "step": 1430
+ },
+ {
+ "epoch": 19.880349344978168,
+ "grad_norm": 0.0076752156019210815,
+ "learning_rate": 0.002061664960736747,
+ "loss": 4.384668827056885,
+ "step": 1431
+ },
+ {
+ "epoch": 19.894323144104803,
+ "grad_norm": 0.008727089501917362,
+ "learning_rate": 0.00206042541103222,
+ "loss": 4.363401412963867,
+ "step": 1432
+ },
+ {
+ "epoch": 19.90829694323144,
+ "grad_norm": 0.007616396062076092,
+ "learning_rate": 0.002059185421888291,
+ "loss": 4.413188934326172,
+ "step": 1433
+ },
+ {
+ "epoch": 19.92227074235808,
+ "grad_norm": 0.008922954089939594,
+ "learning_rate": 0.0020579449943039967,
+ "loss": 4.383417129516602,
+ "step": 1434
+ },
+ {
+ "epoch": 19.936244541484715,
+ "grad_norm": 0.009063739329576492,
+ "learning_rate": 0.002056704129278727,
+ "loss": 4.385283470153809,
+ "step": 1435
+ },
+ {
+ "epoch": 19.950218340611354,
+ "grad_norm": 0.008912574499845505,
+ "learning_rate": 0.002055462827812225,
+ "loss": 4.378478050231934,
+ "step": 1436
+ },
+ {
+ "epoch": 19.964192139737992,
+ "grad_norm": 0.008317525498569012,
+ "learning_rate": 0.0020542210909045844,
+ "loss": 4.367859840393066,
+ "step": 1437
+ },
+ {
+ "epoch": 19.978165938864628,
+ "grad_norm": 0.009347227402031422,
+ "learning_rate": 0.0020529789195562506,
+ "loss": 4.38238000869751,
+ "step": 1438
+ },
+ {
+ "epoch": 19.992139737991266,
+ "grad_norm": 0.011030065827071667,
+ "learning_rate": 0.0020517363147680183,
+ "loss": 4.372624397277832,
+ "step": 1439
+ },
+ {
+ "epoch": 20.0,
+ "grad_norm": 0.011519700288772583,
+ "learning_rate": 0.002050493277541032,
+ "loss": 4.3906121253967285,
+ "step": 1440
+ },
+ {
+ "epoch": 20.0,
+ "eval_loss": 4.641542434692383,
+ "eval_runtime": 51.6342,
+ "eval_samples_per_second": 47.294,
+ "eval_steps_per_second": 1.491,
+ "step": 1440
+ },
+ {
+ "epoch": 20.01397379912664,
+ "grad_norm": 0.009694516658782959,
+ "learning_rate": 0.0020492498088767836,
+ "loss": 4.32808256149292,
+ "step": 1441
+ },
+ {
+ "epoch": 20.027947598253274,
+ "grad_norm": 0.0097952950745821,
+ "learning_rate": 0.002048005909777114,
+ "loss": 4.280096054077148,
+ "step": 1442
+ },
+ {
+ "epoch": 20.041921397379912,
+ "grad_norm": 0.010393359698355198,
+ "learning_rate": 0.0020467615812442106,
+ "loss": 4.317824363708496,
+ "step": 1443
+ },
+ {
+ "epoch": 20.05589519650655,
+ "grad_norm": 0.00977372657507658,
+ "learning_rate": 0.0020455168242806054,
+ "loss": 4.3445329666137695,
+ "step": 1444
+ },
+ {
+ "epoch": 20.069868995633186,
+ "grad_norm": 0.008568719029426575,
+ "learning_rate": 0.0020442716398891775,
+ "loss": 4.318026542663574,
+ "step": 1445
+ },
+ {
+ "epoch": 20.083842794759825,
+ "grad_norm": 0.009732510894536972,
+ "learning_rate": 0.0020430260290731487,
+ "loss": 4.330609321594238,
+ "step": 1446
+ },
+ {
+ "epoch": 20.097816593886463,
+ "grad_norm": 0.010429969057440758,
+ "learning_rate": 0.0020417799928360864,
+ "loss": 4.316794395446777,
+ "step": 1447
+ },
+ {
+ "epoch": 20.111790393013102,
+ "grad_norm": 0.009292383678257465,
+ "learning_rate": 0.002040533532181898,
+ "loss": 4.317420959472656,
+ "step": 1448
+ },
+ {
+ "epoch": 20.125764192139737,
+ "grad_norm": 0.011222455650568008,
+ "learning_rate": 0.0020392866481148356,
+ "loss": 4.292610168457031,
+ "step": 1449
+ },
+ {
+ "epoch": 20.139737991266376,
+ "grad_norm": 0.013591725379228592,
+ "learning_rate": 0.0020380393416394894,
+ "loss": 4.325689315795898,
+ "step": 1450
+ },
+ {
+ "epoch": 20.153711790393015,
+ "grad_norm": 0.013572026044130325,
+ "learning_rate": 0.002036791613760794,
+ "loss": 4.334960460662842,
+ "step": 1451
+ },
+ {
+ "epoch": 20.16768558951965,
+ "grad_norm": 0.012986521236598492,
+ "learning_rate": 0.0020355434654840197,
+ "loss": 4.33829402923584,
+ "step": 1452
+ },
+ {
+ "epoch": 20.18165938864629,
+ "grad_norm": 0.013321424834430218,
+ "learning_rate": 0.0020342948978147777,
+ "loss": 4.2926926612854,
+ "step": 1453
+ },
+ {
+ "epoch": 20.195633187772927,
+ "grad_norm": 0.012307215481996536,
+ "learning_rate": 0.0020330459117590153,
+ "loss": 4.32757568359375,
+ "step": 1454
+ },
+ {
+ "epoch": 20.209606986899562,
+ "grad_norm": 0.012909360229969025,
+ "learning_rate": 0.0020317965083230197,
+ "loss": 4.362889289855957,
+ "step": 1455
+ },
+ {
+ "epoch": 20.2235807860262,
+ "grad_norm": 0.017746610566973686,
+ "learning_rate": 0.0020305466885134115,
+ "loss": 4.3426313400268555,
+ "step": 1456
+ },
+ {
+ "epoch": 20.23755458515284,
+ "grad_norm": 0.01696564070880413,
+ "learning_rate": 0.002029296453337148,
+ "loss": 4.3298516273498535,
+ "step": 1457
+ },
+ {
+ "epoch": 20.251528384279474,
+ "grad_norm": 0.013348255306482315,
+ "learning_rate": 0.002028045803801522,
+ "loss": 4.360220432281494,
+ "step": 1458
+ },
+ {
+ "epoch": 20.265502183406113,
+ "grad_norm": 0.013225145637989044,
+ "learning_rate": 0.002026794740914158,
+ "loss": 4.346250057220459,
+ "step": 1459
+ },
+ {
+ "epoch": 20.27947598253275,
+ "grad_norm": 0.01569020375609398,
+ "learning_rate": 0.002025543265683016,
+ "loss": 4.308025360107422,
+ "step": 1460
+ },
+ {
+ "epoch": 20.29344978165939,
+ "grad_norm": 0.018612321466207504,
+ "learning_rate": 0.002024291379116385,
+ "loss": 4.330190181732178,
+ "step": 1461
+ },
+ {
+ "epoch": 20.307423580786025,
+ "grad_norm": 0.01619943603873253,
+ "learning_rate": 0.00202303908222289,
+ "loss": 4.339552402496338,
+ "step": 1462
+ },
+ {
+ "epoch": 20.321397379912664,
+ "grad_norm": 0.015385213308036327,
+ "learning_rate": 0.0020217863760114826,
+ "loss": 4.3412184715271,
+ "step": 1463
+ },
+ {
+ "epoch": 20.335371179039303,
+ "grad_norm": 0.018517112359404564,
+ "learning_rate": 0.002020533261491446,
+ "loss": 4.3289313316345215,
+ "step": 1464
+ },
+ {
+ "epoch": 20.349344978165938,
+ "grad_norm": 0.014868509024381638,
+ "learning_rate": 0.0020192797396723903,
+ "loss": 4.359846115112305,
+ "step": 1465
+ },
+ {
+ "epoch": 20.363318777292577,
+ "grad_norm": 0.014240248128771782,
+ "learning_rate": 0.0020180258115642578,
+ "loss": 4.369707107543945,
+ "step": 1466
+ },
+ {
+ "epoch": 20.377292576419215,
+ "grad_norm": 0.018113158643245697,
+ "learning_rate": 0.0020167714781773143,
+ "loss": 4.362668514251709,
+ "step": 1467
+ },
+ {
+ "epoch": 20.39126637554585,
+ "grad_norm": 0.01802460104227066,
+ "learning_rate": 0.002015516740522154,
+ "loss": 4.344742298126221,
+ "step": 1468
+ },
+ {
+ "epoch": 20.40524017467249,
+ "grad_norm": 0.01414549071341753,
+ "learning_rate": 0.0020142615996096964,
+ "loss": 4.358080863952637,
+ "step": 1469
+ },
+ {
+ "epoch": 20.419213973799128,
+ "grad_norm": 0.012157700955867767,
+ "learning_rate": 0.002013006056451186,
+ "loss": 4.347763538360596,
+ "step": 1470
+ },
+ {
+ "epoch": 20.433187772925763,
+ "grad_norm": 0.011268534697592258,
+ "learning_rate": 0.0020117501120581905,
+ "loss": 4.335859298706055,
+ "step": 1471
+ },
+ {
+ "epoch": 20.4471615720524,
+ "grad_norm": 0.010395628400146961,
+ "learning_rate": 0.002010493767442602,
+ "loss": 4.341609001159668,
+ "step": 1472
+ },
+ {
+ "epoch": 20.46113537117904,
+ "grad_norm": 0.012147950939834118,
+ "learning_rate": 0.002009237023616635,
+ "loss": 4.328502655029297,
+ "step": 1473
+ },
+ {
+ "epoch": 20.475109170305675,
+ "grad_norm": 0.00945813488215208,
+ "learning_rate": 0.0020079798815928252,
+ "loss": 4.311186790466309,
+ "step": 1474
+ },
+ {
+ "epoch": 20.489082969432314,
+ "grad_norm": 0.009489500895142555,
+ "learning_rate": 0.0020067223423840294,
+ "loss": 4.371470928192139,
+ "step": 1475
+ },
+ {
+ "epoch": 20.503056768558952,
+ "grad_norm": 0.009360029362142086,
+ "learning_rate": 0.0020054644070034234,
+ "loss": 4.33941125869751,
+ "step": 1476
+ },
+ {
+ "epoch": 20.51703056768559,
+ "grad_norm": 0.010032779537141323,
+ "learning_rate": 0.0020042060764645037,
+ "loss": 4.365797519683838,
+ "step": 1477
+ },
+ {
+ "epoch": 20.531004366812226,
+ "grad_norm": 0.009367046877741814,
+ "learning_rate": 0.002002947351781084,
+ "loss": 4.29443359375,
+ "step": 1478
+ },
+ {
+ "epoch": 20.544978165938865,
+ "grad_norm": 0.008369733579456806,
+ "learning_rate": 0.0020016882339672967,
+ "loss": 4.294233322143555,
+ "step": 1479
+ },
+ {
+ "epoch": 20.558951965065503,
+ "grad_norm": 0.010394785553216934,
+ "learning_rate": 0.0020004287240375893,
+ "loss": 4.352355003356934,
+ "step": 1480
+ },
+ {
+ "epoch": 20.57292576419214,
+ "grad_norm": 0.01068816613405943,
+ "learning_rate": 0.001999168823006728,
+ "loss": 4.342516899108887,
+ "step": 1481
+ },
+ {
+ "epoch": 20.586899563318777,
+ "grad_norm": 0.009975973516702652,
+ "learning_rate": 0.0019979085318897905,
+ "loss": 4.346763610839844,
+ "step": 1482
+ },
+ {
+ "epoch": 20.600873362445416,
+ "grad_norm": 0.009618151932954788,
+ "learning_rate": 0.001996647851702171,
+ "loss": 4.329115867614746,
+ "step": 1483
+ },
+ {
+ "epoch": 20.61484716157205,
+ "grad_norm": 0.008399694226682186,
+ "learning_rate": 0.001995386783459577,
+ "loss": 4.28736686706543,
+ "step": 1484
+ },
+ {
+ "epoch": 20.62882096069869,
+ "grad_norm": 0.008235268294811249,
+ "learning_rate": 0.001994125328178029,
+ "loss": 4.339872360229492,
+ "step": 1485
+ },
+ {
+ "epoch": 20.64279475982533,
+ "grad_norm": 0.00836547277867794,
+ "learning_rate": 0.0019928634868738577,
+ "loss": 4.328156471252441,
+ "step": 1486
+ },
+ {
+ "epoch": 20.656768558951963,
+ "grad_norm": 0.009619626216590405,
+ "learning_rate": 0.001991601260563706,
+ "loss": 4.337394714355469,
+ "step": 1487
+ },
+ {
+ "epoch": 20.670742358078602,
+ "grad_norm": 0.009814038872718811,
+ "learning_rate": 0.0019903386502645275,
+ "loss": 4.325926780700684,
+ "step": 1488
+ },
+ {
+ "epoch": 20.68471615720524,
+ "grad_norm": 0.009588954970240593,
+ "learning_rate": 0.001989075656993584,
+ "loss": 4.286477088928223,
+ "step": 1489
+ },
+ {
+ "epoch": 20.69868995633188,
+ "grad_norm": 0.010471943765878677,
+ "learning_rate": 0.001987812281768447,
+ "loss": 4.322466850280762,
+ "step": 1490
+ },
+ {
+ "epoch": 20.712663755458514,
+ "grad_norm": 0.010239035822451115,
+ "learning_rate": 0.001986548525606994,
+ "loss": 4.3487653732299805,
+ "step": 1491
+ },
+ {
+ "epoch": 20.726637554585153,
+ "grad_norm": 0.008935568854212761,
+ "learning_rate": 0.001985284389527411,
+ "loss": 4.303601264953613,
+ "step": 1492
+ },
+ {
+ "epoch": 20.74061135371179,
+ "grad_norm": 0.008624759502708912,
+ "learning_rate": 0.00198401987454819,
+ "loss": 4.34036922454834,
+ "step": 1493
+ },
+ {
+ "epoch": 20.754585152838427,
+ "grad_norm": 0.009296547621488571,
+ "learning_rate": 0.0019827549816881277,
+ "loss": 4.273357391357422,
+ "step": 1494
+ },
+ {
+ "epoch": 20.768558951965066,
+ "grad_norm": 0.009071138687431812,
+ "learning_rate": 0.0019814897119663248,
+ "loss": 4.321051597595215,
+ "step": 1495
+ },
+ {
+ "epoch": 20.782532751091704,
+ "grad_norm": 0.008341389708220959,
+ "learning_rate": 0.0019802240664021876,
+ "loss": 4.343242645263672,
+ "step": 1496
+ },
+ {
+ "epoch": 20.79650655021834,
+ "grad_norm": 0.007881715893745422,
+ "learning_rate": 0.0019789580460154227,
+ "loss": 4.32067346572876,
+ "step": 1497
+ },
+ {
+ "epoch": 20.810480349344978,
+ "grad_norm": 0.008750553242862225,
+ "learning_rate": 0.0019776916518260404,
+ "loss": 4.330263614654541,
+ "step": 1498
+ },
+ {
+ "epoch": 20.824454148471617,
+ "grad_norm": 0.008591460064053535,
+ "learning_rate": 0.001976424884854351,
+ "loss": 4.311957359313965,
+ "step": 1499
+ },
+ {
+ "epoch": 20.83842794759825,
+ "grad_norm": 0.008084543980658054,
+ "learning_rate": 0.001975157746120966,
+ "loss": 4.331319808959961,
+ "step": 1500
+ },
+ {
+ "epoch": 20.85240174672489,
+ "grad_norm": 0.008036594837903976,
+ "learning_rate": 0.001973890236646797,
+ "loss": 4.382656097412109,
+ "step": 1501
+ },
+ {
+ "epoch": 20.86637554585153,
+ "grad_norm": 0.00805772002786398,
+ "learning_rate": 0.0019726223574530527,
+ "loss": 4.297247886657715,
+ "step": 1502
+ },
+ {
+ "epoch": 20.880349344978168,
+ "grad_norm": 0.009019420482218266,
+ "learning_rate": 0.0019713541095612405,
+ "loss": 4.326526165008545,
+ "step": 1503
+ },
+ {
+ "epoch": 20.894323144104803,
+ "grad_norm": 0.009953883476555347,
+ "learning_rate": 0.0019700854939931653,
+ "loss": 4.327356338500977,
+ "step": 1504
+ },
+ {
+ "epoch": 20.90829694323144,
+ "grad_norm": 0.009340492077171803,
+ "learning_rate": 0.0019688165117709273,
+ "loss": 4.343993186950684,
+ "step": 1505
+ },
+ {
+ "epoch": 20.92227074235808,
+ "grad_norm": 0.010395721532404423,
+ "learning_rate": 0.0019675471639169226,
+ "loss": 4.284934997558594,
+ "step": 1506
+ },
+ {
+ "epoch": 20.936244541484715,
+ "grad_norm": 0.009946230798959732,
+ "learning_rate": 0.0019662774514538415,
+ "loss": 4.288936614990234,
+ "step": 1507
+ },
+ {
+ "epoch": 20.950218340611354,
+ "grad_norm": 0.010981100611388683,
+ "learning_rate": 0.0019650073754046695,
+ "loss": 4.339935779571533,
+ "step": 1508
+ },
+ {
+ "epoch": 20.964192139737992,
+ "grad_norm": 0.009688783437013626,
+ "learning_rate": 0.0019637369367926827,
+ "loss": 4.381425857543945,
+ "step": 1509
+ },
+ {
+ "epoch": 20.978165938864628,
+ "grad_norm": 0.009442255832254887,
+ "learning_rate": 0.001962466136641452,
+ "loss": 4.304185390472412,
+ "step": 1510
+ },
+ {
+ "epoch": 20.992139737991266,
+ "grad_norm": 0.009371291846036911,
+ "learning_rate": 0.0019611949759748364,
+ "loss": 4.306120872497559,
+ "step": 1511
+ },
+ {
+ "epoch": 21.0,
+ "grad_norm": 0.009544814005494118,
+ "learning_rate": 0.001959923455816988,
+ "loss": 4.326251983642578,
+ "step": 1512
+ },
+ {
+ "epoch": 21.0,
+ "eval_loss": 4.602745532989502,
+ "eval_runtime": 53.5041,
+ "eval_samples_per_second": 45.641,
+ "eval_steps_per_second": 1.439,
+ "step": 1512
+ },
+ {
+ "epoch": 21.01397379912664,
+ "grad_norm": 0.01010364294052124,
+ "learning_rate": 0.0019586515771923474,
+ "loss": 4.24491024017334,
+ "step": 1513
+ },
+ {
+ "epoch": 21.027947598253274,
+ "grad_norm": 0.012722729705274105,
+ "learning_rate": 0.0019573793411256446,
+ "loss": 4.257739543914795,
+ "step": 1514
+ },
+ {
+ "epoch": 21.041921397379912,
+ "grad_norm": 0.01307204831391573,
+ "learning_rate": 0.0019561067486418967,
+ "loss": 4.291684150695801,
+ "step": 1515
+ },
+ {
+ "epoch": 21.05589519650655,
+ "grad_norm": 0.011215966194868088,
+ "learning_rate": 0.0019548338007664098,
+ "loss": 4.237947463989258,
+ "step": 1516
+ },
+ {
+ "epoch": 21.069868995633186,
+ "grad_norm": 0.012943019159138203,
+ "learning_rate": 0.0019535604985247727,
+ "loss": 4.242859363555908,
+ "step": 1517
+ },
+ {
+ "epoch": 21.083842794759825,
+ "grad_norm": 0.01337282732129097,
+ "learning_rate": 0.0019522868429428636,
+ "loss": 4.238934516906738,
+ "step": 1518
+ },
+ {
+ "epoch": 21.097816593886463,
+ "grad_norm": 0.013697991147637367,
+ "learning_rate": 0.0019510128350468438,
+ "loss": 4.285760402679443,
+ "step": 1519
+ },
+ {
+ "epoch": 21.111790393013102,
+ "grad_norm": 0.012182159349322319,
+ "learning_rate": 0.001949738475863158,
+ "loss": 4.290044784545898,
+ "step": 1520
+ },
+ {
+ "epoch": 21.125764192139737,
+ "grad_norm": 0.012620339170098305,
+ "learning_rate": 0.0019484637664185346,
+ "loss": 4.267187118530273,
+ "step": 1521
+ },
+ {
+ "epoch": 21.139737991266376,
+ "grad_norm": 0.011284958571195602,
+ "learning_rate": 0.0019471887077399837,
+ "loss": 4.26662540435791,
+ "step": 1522
+ },
+ {
+ "epoch": 21.153711790393015,
+ "grad_norm": 0.01161450706422329,
+ "learning_rate": 0.001945913300854797,
+ "loss": 4.283192157745361,
+ "step": 1523
+ },
+ {
+ "epoch": 21.16768558951965,
+ "grad_norm": 0.010022968985140324,
+ "learning_rate": 0.001944637546790548,
+ "loss": 4.295843124389648,
+ "step": 1524
+ },
+ {
+ "epoch": 21.18165938864629,
+ "grad_norm": 0.010341878980398178,
+ "learning_rate": 0.001943361446575087,
+ "loss": 4.295636177062988,
+ "step": 1525
+ },
+ {
+ "epoch": 21.195633187772927,
+ "grad_norm": 0.011549185030162334,
+ "learning_rate": 0.0019420850012365457,
+ "loss": 4.288768768310547,
+ "step": 1526
+ },
+ {
+ "epoch": 21.209606986899562,
+ "grad_norm": 0.012518810108304024,
+ "learning_rate": 0.0019408082118033331,
+ "loss": 4.229338645935059,
+ "step": 1527
+ },
+ {
+ "epoch": 21.2235807860262,
+ "grad_norm": 0.012035328894853592,
+ "learning_rate": 0.0019395310793041353,
+ "loss": 4.258639335632324,
+ "step": 1528
+ },
+ {
+ "epoch": 21.23755458515284,
+ "grad_norm": 0.011193782091140747,
+ "learning_rate": 0.0019382536047679146,
+ "loss": 4.239264488220215,
+ "step": 1529
+ },
+ {
+ "epoch": 21.251528384279474,
+ "grad_norm": 0.009930136613547802,
+ "learning_rate": 0.0019369757892239096,
+ "loss": 4.252286434173584,
+ "step": 1530
+ },
+ {
+ "epoch": 21.265502183406113,
+ "grad_norm": 0.00953229982405901,
+ "learning_rate": 0.0019356976337016327,
+ "loss": 4.293435096740723,
+ "step": 1531
+ },
+ {
+ "epoch": 21.27947598253275,
+ "grad_norm": 0.010618340224027634,
+ "learning_rate": 0.0019344191392308716,
+ "loss": 4.235588550567627,
+ "step": 1532
+ },
+ {
+ "epoch": 21.29344978165939,
+ "grad_norm": 0.008806651458144188,
+ "learning_rate": 0.0019331403068416856,
+ "loss": 4.26716423034668,
+ "step": 1533
+ },
+ {
+ "epoch": 21.307423580786025,
+ "grad_norm": 0.008934796787798405,
+ "learning_rate": 0.0019318611375644065,
+ "loss": 4.258462905883789,
+ "step": 1534
+ },
+ {
+ "epoch": 21.321397379912664,
+ "grad_norm": 0.009283283725380898,
+ "learning_rate": 0.0019305816324296386,
+ "loss": 4.257868766784668,
+ "step": 1535
+ },
+ {
+ "epoch": 21.335371179039303,
+ "grad_norm": 0.009224812500178814,
+ "learning_rate": 0.0019293017924682555,
+ "loss": 4.258486747741699,
+ "step": 1536
+ },
+ {
+ "epoch": 21.349344978165938,
+ "grad_norm": 0.008957969956099987,
+ "learning_rate": 0.0019280216187114016,
+ "loss": 4.309563636779785,
+ "step": 1537
+ },
+ {
+ "epoch": 21.363318777292577,
+ "grad_norm": 0.010284051299095154,
+ "learning_rate": 0.0019267411121904894,
+ "loss": 4.275271415710449,
+ "step": 1538
+ },
+ {
+ "epoch": 21.377292576419215,
+ "grad_norm": 0.009167762473225594,
+ "learning_rate": 0.0019254602739372001,
+ "loss": 4.238452911376953,
+ "step": 1539
+ },
+ {
+ "epoch": 21.39126637554585,
+ "grad_norm": 0.009411636739969254,
+ "learning_rate": 0.0019241791049834816,
+ "loss": 4.2735595703125,
+ "step": 1540
+ },
+ {
+ "epoch": 21.40524017467249,
+ "grad_norm": 0.009044171310961246,
+ "learning_rate": 0.0019228976063615497,
+ "loss": 4.27017879486084,
+ "step": 1541
+ },
+ {
+ "epoch": 21.419213973799128,
+ "grad_norm": 0.008840068243443966,
+ "learning_rate": 0.0019216157791038834,
+ "loss": 4.28773307800293,
+ "step": 1542
+ },
+ {
+ "epoch": 21.433187772925763,
+ "grad_norm": 0.009512359276413918,
+ "learning_rate": 0.001920333624243228,
+ "loss": 4.266008377075195,
+ "step": 1543
+ },
+ {
+ "epoch": 21.4471615720524,
+ "grad_norm": 0.009946667589247227,
+ "learning_rate": 0.001919051142812593,
+ "loss": 4.279062271118164,
+ "step": 1544
+ },
+ {
+ "epoch": 21.46113537117904,
+ "grad_norm": 0.00899894442409277,
+ "learning_rate": 0.0019177683358452505,
+ "loss": 4.277679443359375,
+ "step": 1545
+ },
+ {
+ "epoch": 21.475109170305675,
+ "grad_norm": 0.00909668579697609,
+ "learning_rate": 0.0019164852043747338,
+ "loss": 4.311397552490234,
+ "step": 1546
+ },
+ {
+ "epoch": 21.489082969432314,
+ "grad_norm": 0.009787327609956264,
+ "learning_rate": 0.0019152017494348408,
+ "loss": 4.30875825881958,
+ "step": 1547
+ },
+ {
+ "epoch": 21.503056768558952,
+ "grad_norm": 0.008983899839222431,
+ "learning_rate": 0.001913917972059626,
+ "loss": 4.259649276733398,
+ "step": 1548
+ },
+ {
+ "epoch": 21.51703056768559,
+ "grad_norm": 0.008411582559347153,
+ "learning_rate": 0.0019126338732834071,
+ "loss": 4.259811878204346,
+ "step": 1549
+ },
+ {
+ "epoch": 21.531004366812226,
+ "grad_norm": 0.009162918664515018,
+ "learning_rate": 0.0019113494541407579,
+ "loss": 4.271515846252441,
+ "step": 1550
+ },
+ {
+ "epoch": 21.544978165938865,
+ "grad_norm": 0.011304561980068684,
+ "learning_rate": 0.0019100647156665125,
+ "loss": 4.269217491149902,
+ "step": 1551
+ },
+ {
+ "epoch": 21.558951965065503,
+ "grad_norm": 0.012483394704759121,
+ "learning_rate": 0.0019087796588957617,
+ "loss": 4.2949299812316895,
+ "step": 1552
+ },
+ {
+ "epoch": 21.57292576419214,
+ "grad_norm": 0.012091380544006824,
+ "learning_rate": 0.0019074942848638522,
+ "loss": 4.289752006530762,
+ "step": 1553
+ },
+ {
+ "epoch": 21.586899563318777,
+ "grad_norm": 0.010011567734181881,
+ "learning_rate": 0.0019062085946063869,
+ "loss": 4.294095993041992,
+ "step": 1554
+ },
+ {
+ "epoch": 21.600873362445416,
+ "grad_norm": 0.008766217157244682,
+ "learning_rate": 0.0019049225891592226,
+ "loss": 4.256425857543945,
+ "step": 1555
+ },
+ {
+ "epoch": 21.61484716157205,
+ "grad_norm": 0.012862488627433777,
+ "learning_rate": 0.001903636269558471,
+ "loss": 4.287847995758057,
+ "step": 1556
+ },
+ {
+ "epoch": 21.62882096069869,
+ "grad_norm": 0.013917063362896442,
+ "learning_rate": 0.0019023496368404972,
+ "loss": 4.282784938812256,
+ "step": 1557
+ },
+ {
+ "epoch": 21.64279475982533,
+ "grad_norm": 0.012129023671150208,
+ "learning_rate": 0.0019010626920419173,
+ "loss": 4.2486419677734375,
+ "step": 1558
+ },
+ {
+ "epoch": 21.656768558951963,
+ "grad_norm": 0.010348426178097725,
+ "learning_rate": 0.0018997754361995986,
+ "loss": 4.289836883544922,
+ "step": 1559
+ },
+ {
+ "epoch": 21.670742358078602,
+ "grad_norm": 0.011821693740785122,
+ "learning_rate": 0.001898487870350662,
+ "loss": 4.214344024658203,
+ "step": 1560
+ },
+ {
+ "epoch": 21.68471615720524,
+ "grad_norm": 0.013136724010109901,
+ "learning_rate": 0.0018971999955324742,
+ "loss": 4.24208402633667,
+ "step": 1561
+ },
+ {
+ "epoch": 21.69868995633188,
+ "grad_norm": 0.013173168525099754,
+ "learning_rate": 0.0018959118127826538,
+ "loss": 4.270164489746094,
+ "step": 1562
+ },
+ {
+ "epoch": 21.712663755458514,
+ "grad_norm": 0.011018159799277782,
+ "learning_rate": 0.0018946233231390656,
+ "loss": 4.2571916580200195,
+ "step": 1563
+ },
+ {
+ "epoch": 21.726637554585153,
+ "grad_norm": 0.01018658559769392,
+ "learning_rate": 0.0018933345276398227,
+ "loss": 4.283267498016357,
+ "step": 1564
+ },
+ {
+ "epoch": 21.74061135371179,
+ "grad_norm": 0.009412688203155994,
+ "learning_rate": 0.0018920454273232848,
+ "loss": 4.272198677062988,
+ "step": 1565
+ },
+ {
+ "epoch": 21.754585152838427,
+ "grad_norm": 0.007799267303198576,
+ "learning_rate": 0.0018907560232280557,
+ "loss": 4.260125637054443,
+ "step": 1566
+ },
+ {
+ "epoch": 21.768558951965066,
+ "grad_norm": 0.008747464045882225,
+ "learning_rate": 0.0018894663163929863,
+ "loss": 4.2539215087890625,
+ "step": 1567
+ },
+ {
+ "epoch": 21.782532751091704,
+ "grad_norm": 0.008003434166312218,
+ "learning_rate": 0.0018881763078571689,
+ "loss": 4.245024681091309,
+ "step": 1568
+ },
+ {
+ "epoch": 21.79650655021834,
+ "grad_norm": 0.007593846879899502,
+ "learning_rate": 0.0018868859986599415,
+ "loss": 4.237929344177246,
+ "step": 1569
+ },
+ {
+ "epoch": 21.810480349344978,
+ "grad_norm": 0.008099420927464962,
+ "learning_rate": 0.0018855953898408815,
+ "loss": 4.2789106369018555,
+ "step": 1570
+ },
+ {
+ "epoch": 21.824454148471617,
+ "grad_norm": 0.008270486257970333,
+ "learning_rate": 0.0018843044824398097,
+ "loss": 4.263522148132324,
+ "step": 1571
+ },
+ {
+ "epoch": 21.83842794759825,
+ "grad_norm": 0.009260379709303379,
+ "learning_rate": 0.0018830132774967866,
+ "loss": 4.275835037231445,
+ "step": 1572
+ },
+ {
+ "epoch": 21.85240174672489,
+ "grad_norm": 0.009289492852985859,
+ "learning_rate": 0.0018817217760521132,
+ "loss": 4.313778400421143,
+ "step": 1573
+ },
+ {
+ "epoch": 21.86637554585153,
+ "grad_norm": 0.009493921883404255,
+ "learning_rate": 0.0018804299791463281,
+ "loss": 4.303672790527344,
+ "step": 1574
+ },
+ {
+ "epoch": 21.880349344978168,
+ "grad_norm": 0.009020023047924042,
+ "learning_rate": 0.0018791378878202105,
+ "loss": 4.245473384857178,
+ "step": 1575
+ },
+ {
+ "epoch": 21.894323144104803,
+ "grad_norm": 0.009446556679904461,
+ "learning_rate": 0.0018778455031147724,
+ "loss": 4.267327785491943,
+ "step": 1576
+ },
+ {
+ "epoch": 21.90829694323144,
+ "grad_norm": 0.009499974548816681,
+ "learning_rate": 0.0018765528260712672,
+ "loss": 4.301205635070801,
+ "step": 1577
+ },
+ {
+ "epoch": 21.92227074235808,
+ "grad_norm": 0.008399393409490585,
+ "learning_rate": 0.00187525985773118,
+ "loss": 4.307802677154541,
+ "step": 1578
+ },
+ {
+ "epoch": 21.936244541484715,
+ "grad_norm": 0.009519701823592186,
+ "learning_rate": 0.0018739665991362327,
+ "loss": 4.284796714782715,
+ "step": 1579
+ },
+ {
+ "epoch": 21.950218340611354,
+ "grad_norm": 0.01125232595950365,
+ "learning_rate": 0.0018726730513283804,
+ "loss": 4.285930633544922,
+ "step": 1580
+ },
+ {
+ "epoch": 21.964192139737992,
+ "grad_norm": 0.011608537286520004,
+ "learning_rate": 0.00187137921534981,
+ "loss": 4.257689476013184,
+ "step": 1581
+ },
+ {
+ "epoch": 21.978165938864628,
+ "grad_norm": 0.01064533181488514,
+ "learning_rate": 0.001870085092242944,
+ "loss": 4.249063968658447,
+ "step": 1582
+ },
+ {
+ "epoch": 21.992139737991266,
+ "grad_norm": 0.009854599833488464,
+ "learning_rate": 0.0018687906830504321,
+ "loss": 4.255159378051758,
+ "step": 1583
+ },
+ {
+ "epoch": 22.0,
+ "grad_norm": 0.01121739111840725,
+ "learning_rate": 0.0018674959888151576,
+ "loss": 4.287777900695801,
+ "step": 1584
+ },
+ {
+ "epoch": 22.0,
+ "eval_loss": 4.5675368309021,
+ "eval_runtime": 51.8555,
+ "eval_samples_per_second": 47.092,
+ "eval_steps_per_second": 1.485,
+ "step": 1584
+ },
+ {
+ "epoch": 22.01397379912664,
+ "grad_norm": 0.0104464590549469,
+ "learning_rate": 0.0018662010105802314,
+ "loss": 4.206080436706543,
+ "step": 1585
+ },
+ {
+ "epoch": 22.027947598253274,
+ "grad_norm": 0.012800181284546852,
+ "learning_rate": 0.0018649057493889945,
+ "loss": 4.213550567626953,
+ "step": 1586
+ },
+ {
+ "epoch": 22.041921397379912,
+ "grad_norm": 0.013382901437580585,
+ "learning_rate": 0.001863610206285015,
+ "loss": 4.190858840942383,
+ "step": 1587
+ },
+ {
+ "epoch": 22.05589519650655,
+ "grad_norm": 0.014347050338983536,
+ "learning_rate": 0.0018623143823120894,
+ "loss": 4.22884464263916,
+ "step": 1588
+ },
+ {
+ "epoch": 22.069868995633186,
+ "grad_norm": 0.01347152329981327,
+ "learning_rate": 0.001861018278514239,
+ "loss": 4.215579032897949,
+ "step": 1589
+ },
+ {
+ "epoch": 22.083842794759825,
+ "grad_norm": 0.012786377221345901,
+ "learning_rate": 0.0018597218959357112,
+ "loss": 4.174999237060547,
+ "step": 1590
+ },
+ {
+ "epoch": 22.097816593886463,
+ "grad_norm": 0.014137202873826027,
+ "learning_rate": 0.001858425235620979,
+ "loss": 4.21709680557251,
+ "step": 1591
+ },
+ {
+ "epoch": 22.111790393013102,
+ "grad_norm": 0.013864399865269661,
+ "learning_rate": 0.001857128298614737,
+ "loss": 4.2375030517578125,
+ "step": 1592
+ },
+ {
+ "epoch": 22.125764192139737,
+ "grad_norm": 0.010397658683359623,
+ "learning_rate": 0.0018558310859619048,
+ "loss": 4.231938362121582,
+ "step": 1593
+ },
+ {
+ "epoch": 22.139737991266376,
+ "grad_norm": 0.00888819433748722,
+ "learning_rate": 0.0018545335987076223,
+ "loss": 4.195562839508057,
+ "step": 1594
+ },
+ {
+ "epoch": 22.153711790393015,
+ "grad_norm": 0.010698067955672741,
+ "learning_rate": 0.001853235837897252,
+ "loss": 4.22847318649292,
+ "step": 1595
+ },
+ {
+ "epoch": 22.16768558951965,
+ "grad_norm": 0.00987589918076992,
+ "learning_rate": 0.0018519378045763766,
+ "loss": 4.2147040367126465,
+ "step": 1596
+ },
+ {
+ "epoch": 22.18165938864629,
+ "grad_norm": 0.011451782658696175,
+ "learning_rate": 0.001850639499790798,
+ "loss": 4.179263114929199,
+ "step": 1597
+ },
+ {
+ "epoch": 22.195633187772927,
+ "grad_norm": 0.011224704794585705,
+ "learning_rate": 0.0018493409245865368,
+ "loss": 4.23030948638916,
+ "step": 1598
+ },
+ {
+ "epoch": 22.209606986899562,
+ "grad_norm": 0.00996861606836319,
+ "learning_rate": 0.0018480420800098313,
+ "loss": 4.194283962249756,
+ "step": 1599
+ },
+ {
+ "epoch": 22.2235807860262,
+ "grad_norm": 0.012143119238317013,
+ "learning_rate": 0.0018467429671071374,
+ "loss": 4.221704959869385,
+ "step": 1600
+ },
+ {
+ "epoch": 22.23755458515284,
+ "grad_norm": 0.013126783072948456,
+ "learning_rate": 0.0018454435869251268,
+ "loss": 4.187748908996582,
+ "step": 1601
+ },
+ {
+ "epoch": 22.251528384279474,
+ "grad_norm": 0.013031953014433384,
+ "learning_rate": 0.001844143940510687,
+ "loss": 4.248250961303711,
+ "step": 1602
+ },
+ {
+ "epoch": 22.265502183406113,
+ "grad_norm": 0.012868993915617466,
+ "learning_rate": 0.001842844028910919,
+ "loss": 4.20161247253418,
+ "step": 1603
+ },
+ {
+ "epoch": 22.27947598253275,
+ "grad_norm": 0.010102041065692902,
+ "learning_rate": 0.0018415438531731395,
+ "loss": 4.213654518127441,
+ "step": 1604
+ },
+ {
+ "epoch": 22.29344978165939,
+ "grad_norm": 0.009599820710718632,
+ "learning_rate": 0.0018402434143448747,
+ "loss": 4.206313133239746,
+ "step": 1605
+ },
+ {
+ "epoch": 22.307423580786025,
+ "grad_norm": 0.011536866426467896,
+ "learning_rate": 0.0018389427134738658,
+ "loss": 4.237072944641113,
+ "step": 1606
+ },
+ {
+ "epoch": 22.321397379912664,
+ "grad_norm": 0.009558089077472687,
+ "learning_rate": 0.0018376417516080636,
+ "loss": 4.182195663452148,
+ "step": 1607
+ },
+ {
+ "epoch": 22.335371179039303,
+ "grad_norm": 0.01056219544261694,
+ "learning_rate": 0.0018363405297956302,
+ "loss": 4.232585906982422,
+ "step": 1608
+ },
+ {
+ "epoch": 22.349344978165938,
+ "grad_norm": 0.012673584744334221,
+ "learning_rate": 0.0018350390490849354,
+ "loss": 4.228830337524414,
+ "step": 1609
+ },
+ {
+ "epoch": 22.363318777292577,
+ "grad_norm": 0.011215884238481522,
+ "learning_rate": 0.0018337373105245598,
+ "loss": 4.216056823730469,
+ "step": 1610
+ },
+ {
+ "epoch": 22.377292576419215,
+ "grad_norm": 0.011281915940344334,
+ "learning_rate": 0.00183243531516329,
+ "loss": 4.233404159545898,
+ "step": 1611
+ },
+ {
+ "epoch": 22.39126637554585,
+ "grad_norm": 0.010381227359175682,
+ "learning_rate": 0.0018311330640501204,
+ "loss": 4.182013034820557,
+ "step": 1612
+ },
+ {
+ "epoch": 22.40524017467249,
+ "grad_norm": 0.00900198519229889,
+ "learning_rate": 0.0018298305582342513,
+ "loss": 4.215488433837891,
+ "step": 1613
+ },
+ {
+ "epoch": 22.419213973799128,
+ "grad_norm": 0.010198601521551609,
+ "learning_rate": 0.0018285277987650876,
+ "loss": 4.19389533996582,
+ "step": 1614
+ },
+ {
+ "epoch": 22.433187772925763,
+ "grad_norm": 0.009965402074158192,
+ "learning_rate": 0.0018272247866922401,
+ "loss": 4.218005657196045,
+ "step": 1615
+ },
+ {
+ "epoch": 22.4471615720524,
+ "grad_norm": 0.009299935773015022,
+ "learning_rate": 0.0018259215230655202,
+ "loss": 4.223684310913086,
+ "step": 1616
+ },
+ {
+ "epoch": 22.46113537117904,
+ "grad_norm": 0.008977951481938362,
+ "learning_rate": 0.0018246180089349462,
+ "loss": 4.215778350830078,
+ "step": 1617
+ },
+ {
+ "epoch": 22.475109170305675,
+ "grad_norm": 0.009261307306587696,
+ "learning_rate": 0.0018233142453507344,
+ "loss": 4.215066909790039,
+ "step": 1618
+ },
+ {
+ "epoch": 22.489082969432314,
+ "grad_norm": 0.008322088047862053,
+ "learning_rate": 0.0018220102333633044,
+ "loss": 4.2188005447387695,
+ "step": 1619
+ },
+ {
+ "epoch": 22.503056768558952,
+ "grad_norm": 0.008716144599020481,
+ "learning_rate": 0.0018207059740232739,
+ "loss": 4.187867641448975,
+ "step": 1620
+ },
+ {
+ "epoch": 22.51703056768559,
+ "grad_norm": 0.008897020481526852,
+ "learning_rate": 0.0018194014683814625,
+ "loss": 4.218034267425537,
+ "step": 1621
+ },
+ {
+ "epoch": 22.531004366812226,
+ "grad_norm": 0.007298174314200878,
+ "learning_rate": 0.0018180967174888857,
+ "loss": 4.2017011642456055,
+ "step": 1622
+ },
+ {
+ "epoch": 22.544978165938865,
+ "grad_norm": 0.008491789922118187,
+ "learning_rate": 0.0018167917223967582,
+ "loss": 4.239873886108398,
+ "step": 1623
+ },
+ {
+ "epoch": 22.558951965065503,
+ "grad_norm": 0.008413378149271011,
+ "learning_rate": 0.0018154864841564908,
+ "loss": 4.234771251678467,
+ "step": 1624
+ },
+ {
+ "epoch": 22.57292576419214,
+ "grad_norm": 0.008141471073031425,
+ "learning_rate": 0.0018141810038196909,
+ "loss": 4.185944080352783,
+ "step": 1625
+ },
+ {
+ "epoch": 22.586899563318777,
+ "grad_norm": 0.009480944834649563,
+ "learning_rate": 0.0018128752824381603,
+ "loss": 4.206586837768555,
+ "step": 1626
+ },
+ {
+ "epoch": 22.600873362445416,
+ "grad_norm": 0.0106047997251153,
+ "learning_rate": 0.001811569321063894,
+ "loss": 4.235287666320801,
+ "step": 1627
+ },
+ {
+ "epoch": 22.61484716157205,
+ "grad_norm": 0.010663514956831932,
+ "learning_rate": 0.0018102631207490838,
+ "loss": 4.204899311065674,
+ "step": 1628
+ },
+ {
+ "epoch": 22.62882096069869,
+ "grad_norm": 0.011473523452877998,
+ "learning_rate": 0.001808956682546109,
+ "loss": 4.253668308258057,
+ "step": 1629
+ },
+ {
+ "epoch": 22.64279475982533,
+ "grad_norm": 0.01165992021560669,
+ "learning_rate": 0.0018076500075075456,
+ "loss": 4.209845542907715,
+ "step": 1630
+ },
+ {
+ "epoch": 22.656768558951963,
+ "grad_norm": 0.011692410334944725,
+ "learning_rate": 0.0018063430966861564,
+ "loss": 4.250368595123291,
+ "step": 1631
+ },
+ {
+ "epoch": 22.670742358078602,
+ "grad_norm": 0.01061506848782301,
+ "learning_rate": 0.0018050359511348965,
+ "loss": 4.193821907043457,
+ "step": 1632
+ },
+ {
+ "epoch": 22.68471615720524,
+ "grad_norm": 0.010510941036045551,
+ "learning_rate": 0.0018037285719069097,
+ "loss": 4.254327774047852,
+ "step": 1633
+ },
+ {
+ "epoch": 22.69868995633188,
+ "grad_norm": 0.011587041430175304,
+ "learning_rate": 0.001802420960055528,
+ "loss": 4.2838945388793945,
+ "step": 1634
+ },
+ {
+ "epoch": 22.712663755458514,
+ "grad_norm": 0.01257420051842928,
+ "learning_rate": 0.0018011131166342693,
+ "loss": 4.258241653442383,
+ "step": 1635
+ },
+ {
+ "epoch": 22.726637554585153,
+ "grad_norm": 0.00936233438551426,
+ "learning_rate": 0.0017998050426968407,
+ "loss": 4.201854705810547,
+ "step": 1636
+ },
+ {
+ "epoch": 22.74061135371179,
+ "grad_norm": 0.010093127377331257,
+ "learning_rate": 0.0017984967392971338,
+ "loss": 4.193051338195801,
+ "step": 1637
+ },
+ {
+ "epoch": 22.754585152838427,
+ "grad_norm": 0.010032009333372116,
+ "learning_rate": 0.0017971882074892236,
+ "loss": 4.230212211608887,
+ "step": 1638
+ },
+ {
+ "epoch": 22.768558951965066,
+ "grad_norm": 0.009455587714910507,
+ "learning_rate": 0.0017958794483273726,
+ "loss": 4.2157673835754395,
+ "step": 1639
+ },
+ {
+ "epoch": 22.782532751091704,
+ "grad_norm": 0.009549238719046116,
+ "learning_rate": 0.001794570462866023,
+ "loss": 4.260272026062012,
+ "step": 1640
+ },
+ {
+ "epoch": 22.79650655021834,
+ "grad_norm": 0.011059699580073357,
+ "learning_rate": 0.0017932612521598013,
+ "loss": 4.214818000793457,
+ "step": 1641
+ },
+ {
+ "epoch": 22.810480349344978,
+ "grad_norm": 0.011653627268970013,
+ "learning_rate": 0.0017919518172635146,
+ "loss": 4.205112457275391,
+ "step": 1642
+ },
+ {
+ "epoch": 22.824454148471617,
+ "grad_norm": 0.012438314035534859,
+ "learning_rate": 0.001790642159232151,
+ "loss": 4.179045677185059,
+ "step": 1643
+ },
+ {
+ "epoch": 22.83842794759825,
+ "grad_norm": 0.011234673671424389,
+ "learning_rate": 0.0017893322791208784,
+ "loss": 4.239317417144775,
+ "step": 1644
+ },
+ {
+ "epoch": 22.85240174672489,
+ "grad_norm": 0.010923845693469048,
+ "learning_rate": 0.0017880221779850434,
+ "loss": 4.230571269989014,
+ "step": 1645
+ },
+ {
+ "epoch": 22.86637554585153,
+ "grad_norm": 0.009219309315085411,
+ "learning_rate": 0.001786711856880171,
+ "loss": 4.229068756103516,
+ "step": 1646
+ },
+ {
+ "epoch": 22.880349344978168,
+ "grad_norm": 0.009849398396909237,
+ "learning_rate": 0.0017854013168619638,
+ "loss": 4.254772186279297,
+ "step": 1647
+ },
+ {
+ "epoch": 22.894323144104803,
+ "grad_norm": 0.01117763202637434,
+ "learning_rate": 0.001784090558986299,
+ "loss": 4.252647399902344,
+ "step": 1648
+ },
+ {
+ "epoch": 22.90829694323144,
+ "grad_norm": 0.010664612054824829,
+ "learning_rate": 0.0017827795843092316,
+ "loss": 4.204596519470215,
+ "step": 1649
+ },
+ {
+ "epoch": 22.92227074235808,
+ "grad_norm": 0.00847950391471386,
+ "learning_rate": 0.0017814683938869895,
+ "loss": 4.211061954498291,
+ "step": 1650
+ },
+ {
+ "epoch": 22.936244541484715,
+ "grad_norm": 0.007305726408958435,
+ "learning_rate": 0.0017801569887759753,
+ "loss": 4.181221008300781,
+ "step": 1651
+ },
+ {
+ "epoch": 22.950218340611354,
+ "grad_norm": 0.007970666512846947,
+ "learning_rate": 0.0017788453700327653,
+ "loss": 4.237855434417725,
+ "step": 1652
+ },
+ {
+ "epoch": 22.964192139737992,
+ "grad_norm": 0.008873614482581615,
+ "learning_rate": 0.0017775335387141054,
+ "loss": 4.229188919067383,
+ "step": 1653
+ },
+ {
+ "epoch": 22.978165938864628,
+ "grad_norm": 0.009691286832094193,
+ "learning_rate": 0.0017762214958769156,
+ "loss": 4.205491542816162,
+ "step": 1654
+ },
+ {
+ "epoch": 22.992139737991266,
+ "grad_norm": 0.009360672906041145,
+ "learning_rate": 0.0017749092425782852,
+ "loss": 4.233466625213623,
+ "step": 1655
+ },
+ {
+ "epoch": 23.0,
+ "grad_norm": 0.009895884431898594,
+ "learning_rate": 0.0017735967798754723,
+ "loss": 4.187763214111328,
+ "step": 1656
+ },
+ {
+ "epoch": 23.0,
+ "eval_loss": 4.535672187805176,
+ "eval_runtime": 53.1736,
+ "eval_samples_per_second": 45.925,
+ "eval_steps_per_second": 1.448,
+ "step": 1656
+ },
+ {
+ "epoch": 23.01397379912664,
+ "grad_norm": 0.00966129545122385,
+ "learning_rate": 0.0017722841088259045,
+ "loss": 4.1451735496521,
+ "step": 1657
+ },
+ {
+ "epoch": 23.027947598253274,
+ "grad_norm": 0.010733380913734436,
+ "learning_rate": 0.0017709712304871774,
+ "loss": 4.1616363525390625,
+ "step": 1658
+ },
+ {
+ "epoch": 23.041921397379912,
+ "grad_norm": 0.0117328567430377,
+ "learning_rate": 0.0017696581459170535,
+ "loss": 4.126331806182861,
+ "step": 1659
+ },
+ {
+ "epoch": 23.05589519650655,
+ "grad_norm": 0.012640298344194889,
+ "learning_rate": 0.0017683448561734607,
+ "loss": 4.163730621337891,
+ "step": 1660
+ },
+ {
+ "epoch": 23.069868995633186,
+ "grad_norm": 0.013074076734483242,
+ "learning_rate": 0.0017670313623144934,
+ "loss": 4.199502944946289,
+ "step": 1661
+ },
+ {
+ "epoch": 23.083842794759825,
+ "grad_norm": 0.01132719125598669,
+ "learning_rate": 0.0017657176653984097,
+ "loss": 4.143380165100098,
+ "step": 1662
+ },
+ {
+ "epoch": 23.097816593886463,
+ "grad_norm": 0.010008333250880241,
+ "learning_rate": 0.0017644037664836319,
+ "loss": 4.170320510864258,
+ "step": 1663
+ },
+ {
+ "epoch": 23.111790393013102,
+ "grad_norm": 0.011090578511357307,
+ "learning_rate": 0.0017630896666287438,
+ "loss": 4.1571478843688965,
+ "step": 1664
+ },
+ {
+ "epoch": 23.125764192139737,
+ "grad_norm": 0.011047456413507462,
+ "learning_rate": 0.0017617753668924924,
+ "loss": 4.101892948150635,
+ "step": 1665
+ },
+ {
+ "epoch": 23.139737991266376,
+ "grad_norm": 0.010098653845489025,
+ "learning_rate": 0.0017604608683337857,
+ "loss": 4.154077529907227,
+ "step": 1666
+ },
+ {
+ "epoch": 23.153711790393015,
+ "grad_norm": 0.010499860160052776,
+ "learning_rate": 0.00175914617201169,
+ "loss": 4.116868019104004,
+ "step": 1667
+ },
+ {
+ "epoch": 23.16768558951965,
+ "grad_norm": 0.010939318686723709,
+ "learning_rate": 0.0017578312789854338,
+ "loss": 4.159144878387451,
+ "step": 1668
+ },
+ {
+ "epoch": 23.18165938864629,
+ "grad_norm": 0.010622999630868435,
+ "learning_rate": 0.0017565161903144023,
+ "loss": 4.171411991119385,
+ "step": 1669
+ },
+ {
+ "epoch": 23.195633187772927,
+ "grad_norm": 0.0108700692653656,
+ "learning_rate": 0.0017552009070581386,
+ "loss": 4.100748062133789,
+ "step": 1670
+ },
+ {
+ "epoch": 23.209606986899562,
+ "grad_norm": 0.009794988669455051,
+ "learning_rate": 0.0017538854302763423,
+ "loss": 4.185161590576172,
+ "step": 1671
+ },
+ {
+ "epoch": 23.2235807860262,
+ "grad_norm": 0.007750354241579771,
+ "learning_rate": 0.0017525697610288702,
+ "loss": 4.149083137512207,
+ "step": 1672
+ },
+ {
+ "epoch": 23.23755458515284,
+ "grad_norm": 0.009923134930431843,
+ "learning_rate": 0.001751253900375732,
+ "loss": 4.205166816711426,
+ "step": 1673
+ },
+ {
+ "epoch": 23.251528384279474,
+ "grad_norm": 0.010811141692101955,
+ "learning_rate": 0.0017499378493770936,
+ "loss": 4.131865978240967,
+ "step": 1674
+ },
+ {
+ "epoch": 23.265502183406113,
+ "grad_norm": 0.010868688113987446,
+ "learning_rate": 0.0017486216090932733,
+ "loss": 4.148791790008545,
+ "step": 1675
+ },
+ {
+ "epoch": 23.27947598253275,
+ "grad_norm": 0.011224126443266869,
+ "learning_rate": 0.001747305180584743,
+ "loss": 4.152503490447998,
+ "step": 1676
+ },
+ {
+ "epoch": 23.29344978165939,
+ "grad_norm": 0.013262554071843624,
+ "learning_rate": 0.001745988564912124,
+ "loss": 4.164926528930664,
+ "step": 1677
+ },
+ {
+ "epoch": 23.307423580786025,
+ "grad_norm": 0.013089599087834358,
+ "learning_rate": 0.001744671763136191,
+ "loss": 4.1626458168029785,
+ "step": 1678
+ },
+ {
+ "epoch": 23.321397379912664,
+ "grad_norm": 0.01327646616846323,
+ "learning_rate": 0.0017433547763178665,
+ "loss": 4.189356803894043,
+ "step": 1679
+ },
+ {
+ "epoch": 23.335371179039303,
+ "grad_norm": 0.014539042487740517,
+ "learning_rate": 0.0017420376055182233,
+ "loss": 4.176387786865234,
+ "step": 1680
+ },
+ {
+ "epoch": 23.349344978165938,
+ "grad_norm": 0.013269098475575447,
+ "learning_rate": 0.001740720251798482,
+ "loss": 4.180716514587402,
+ "step": 1681
+ },
+ {
+ "epoch": 23.363318777292577,
+ "grad_norm": 0.010922391898930073,
+ "learning_rate": 0.0017394027162200105,
+ "loss": 4.163928031921387,
+ "step": 1682
+ },
+ {
+ "epoch": 23.377292576419215,
+ "grad_norm": 0.013355271890759468,
+ "learning_rate": 0.0017380849998443243,
+ "loss": 4.150012016296387,
+ "step": 1683
+ },
+ {
+ "epoch": 23.39126637554585,
+ "grad_norm": 0.014149201102554798,
+ "learning_rate": 0.0017367671037330828,
+ "loss": 4.173713207244873,
+ "step": 1684
+ },
+ {
+ "epoch": 23.40524017467249,
+ "grad_norm": 0.01354879979044199,
+ "learning_rate": 0.0017354490289480918,
+ "loss": 4.150563716888428,
+ "step": 1685
+ },
+ {
+ "epoch": 23.419213973799128,
+ "grad_norm": 0.013788704760372639,
+ "learning_rate": 0.0017341307765512998,
+ "loss": 4.213858604431152,
+ "step": 1686
+ },
+ {
+ "epoch": 23.433187772925763,
+ "grad_norm": 0.011597222648561,
+ "learning_rate": 0.0017328123476047994,
+ "loss": 4.166989326477051,
+ "step": 1687
+ },
+ {
+ "epoch": 23.4471615720524,
+ "grad_norm": 0.01055675558745861,
+ "learning_rate": 0.001731493743170824,
+ "loss": 4.172226428985596,
+ "step": 1688
+ },
+ {
+ "epoch": 23.46113537117904,
+ "grad_norm": 0.011600397527217865,
+ "learning_rate": 0.0017301749643117506,
+ "loss": 4.1656951904296875,
+ "step": 1689
+ },
+ {
+ "epoch": 23.475109170305675,
+ "grad_norm": 0.01114651933312416,
+ "learning_rate": 0.0017288560120900943,
+ "loss": 4.189825057983398,
+ "step": 1690
+ },
+ {
+ "epoch": 23.489082969432314,
+ "grad_norm": 0.009811583906412125,
+ "learning_rate": 0.0017275368875685126,
+ "loss": 4.181164741516113,
+ "step": 1691
+ },
+ {
+ "epoch": 23.503056768558952,
+ "grad_norm": 0.009758968837559223,
+ "learning_rate": 0.0017262175918097992,
+ "loss": 4.163188934326172,
+ "step": 1692
+ },
+ {
+ "epoch": 23.51703056768559,
+ "grad_norm": 0.00999707542359829,
+ "learning_rate": 0.001724898125876887,
+ "loss": 4.184117317199707,
+ "step": 1693
+ },
+ {
+ "epoch": 23.531004366812226,
+ "grad_norm": 0.009528180584311485,
+ "learning_rate": 0.001723578490832846,
+ "loss": 4.16407585144043,
+ "step": 1694
+ },
+ {
+ "epoch": 23.544978165938865,
+ "grad_norm": 0.009717358276247978,
+ "learning_rate": 0.0017222586877408826,
+ "loss": 4.209892272949219,
+ "step": 1695
+ },
+ {
+ "epoch": 23.558951965065503,
+ "grad_norm": 0.010909050703048706,
+ "learning_rate": 0.0017209387176643377,
+ "loss": 4.19977331161499,
+ "step": 1696
+ },
+ {
+ "epoch": 23.57292576419214,
+ "grad_norm": 0.011234750039875507,
+ "learning_rate": 0.0017196185816666875,
+ "loss": 4.19363260269165,
+ "step": 1697
+ },
+ {
+ "epoch": 23.586899563318777,
+ "grad_norm": 0.010300721041858196,
+ "learning_rate": 0.0017182982808115425,
+ "loss": 4.142427921295166,
+ "step": 1698
+ },
+ {
+ "epoch": 23.600873362445416,
+ "grad_norm": 0.00941646471619606,
+ "learning_rate": 0.001716977816162644,
+ "loss": 4.159372329711914,
+ "step": 1699
+ },
+ {
+ "epoch": 23.61484716157205,
+ "grad_norm": 0.009304054081439972,
+ "learning_rate": 0.0017156571887838679,
+ "loss": 4.207922458648682,
+ "step": 1700
+ },
+ {
+ "epoch": 23.62882096069869,
+ "grad_norm": 0.009496180340647697,
+ "learning_rate": 0.0017143363997392184,
+ "loss": 4.137894630432129,
+ "step": 1701
+ },
+ {
+ "epoch": 23.64279475982533,
+ "grad_norm": 0.009257436729967594,
+ "learning_rate": 0.0017130154500928326,
+ "loss": 4.198753833770752,
+ "step": 1702
+ },
+ {
+ "epoch": 23.656768558951963,
+ "grad_norm": 0.00900586973875761,
+ "learning_rate": 0.0017116943409089746,
+ "loss": 4.160684585571289,
+ "step": 1703
+ },
+ {
+ "epoch": 23.670742358078602,
+ "grad_norm": 0.009398494847118855,
+ "learning_rate": 0.00171037307325204,
+ "loss": 4.1819353103637695,
+ "step": 1704
+ },
+ {
+ "epoch": 23.68471615720524,
+ "grad_norm": 0.008633013814687729,
+ "learning_rate": 0.0017090516481865483,
+ "loss": 4.176833152770996,
+ "step": 1705
+ },
+ {
+ "epoch": 23.69868995633188,
+ "grad_norm": 0.00898425467312336,
+ "learning_rate": 0.0017077300667771484,
+ "loss": 4.190657138824463,
+ "step": 1706
+ },
+ {
+ "epoch": 23.712663755458514,
+ "grad_norm": 0.008182398043572903,
+ "learning_rate": 0.0017064083300886155,
+ "loss": 4.173429489135742,
+ "step": 1707
+ },
+ {
+ "epoch": 23.726637554585153,
+ "grad_norm": 0.009589143097400665,
+ "learning_rate": 0.0017050864391858478,
+ "loss": 4.1559014320373535,
+ "step": 1708
+ },
+ {
+ "epoch": 23.74061135371179,
+ "grad_norm": 0.010304383002221584,
+ "learning_rate": 0.0017037643951338698,
+ "loss": 4.155453681945801,
+ "step": 1709
+ },
+ {
+ "epoch": 23.754585152838427,
+ "grad_norm": 0.009943748824298382,
+ "learning_rate": 0.0017024421989978275,
+ "loss": 4.119734287261963,
+ "step": 1710
+ },
+ {
+ "epoch": 23.768558951965066,
+ "grad_norm": 0.009238033555448055,
+ "learning_rate": 0.001701119851842992,
+ "loss": 4.188974857330322,
+ "step": 1711
+ },
+ {
+ "epoch": 23.782532751091704,
+ "grad_norm": 0.008058690465986729,
+ "learning_rate": 0.0016997973547347531,
+ "loss": 4.19934606552124,
+ "step": 1712
+ },
+ {
+ "epoch": 23.79650655021834,
+ "grad_norm": 0.008485953323543072,
+ "learning_rate": 0.0016984747087386233,
+ "loss": 4.165615558624268,
+ "step": 1713
+ },
+ {
+ "epoch": 23.810480349344978,
+ "grad_norm": 0.008476605638861656,
+ "learning_rate": 0.0016971519149202347,
+ "loss": 4.21021842956543,
+ "step": 1714
+ },
+ {
+ "epoch": 23.824454148471617,
+ "grad_norm": 0.008588708005845547,
+ "learning_rate": 0.0016958289743453385,
+ "loss": 4.208064556121826,
+ "step": 1715
+ },
+ {
+ "epoch": 23.83842794759825,
+ "grad_norm": 0.009223178029060364,
+ "learning_rate": 0.0016945058880798037,
+ "loss": 4.158413887023926,
+ "step": 1716
+ },
+ {
+ "epoch": 23.85240174672489,
+ "grad_norm": 0.009196167811751366,
+ "learning_rate": 0.0016931826571896166,
+ "loss": 4.203187465667725,
+ "step": 1717
+ },
+ {
+ "epoch": 23.86637554585153,
+ "grad_norm": 0.009488921612501144,
+ "learning_rate": 0.001691859282740882,
+ "loss": 4.178227424621582,
+ "step": 1718
+ },
+ {
+ "epoch": 23.880349344978168,
+ "grad_norm": 0.009253161028027534,
+ "learning_rate": 0.0016905357657998167,
+ "loss": 4.131230354309082,
+ "step": 1719
+ },
+ {
+ "epoch": 23.894323144104803,
+ "grad_norm": 0.010069048963487148,
+ "learning_rate": 0.0016892121074327559,
+ "loss": 4.210734844207764,
+ "step": 1720
+ },
+ {
+ "epoch": 23.90829694323144,
+ "grad_norm": 0.009945095516741276,
+ "learning_rate": 0.001687888308706147,
+ "loss": 4.184042930603027,
+ "step": 1721
+ },
+ {
+ "epoch": 23.92227074235808,
+ "grad_norm": 0.010483640246093273,
+ "learning_rate": 0.0016865643706865505,
+ "loss": 4.156919479370117,
+ "step": 1722
+ },
+ {
+ "epoch": 23.936244541484715,
+ "grad_norm": 0.011560715734958649,
+ "learning_rate": 0.0016852402944406391,
+ "loss": 4.185520172119141,
+ "step": 1723
+ },
+ {
+ "epoch": 23.950218340611354,
+ "grad_norm": 0.010001277551054955,
+ "learning_rate": 0.0016839160810351978,
+ "loss": 4.226828575134277,
+ "step": 1724
+ },
+ {
+ "epoch": 23.964192139737992,
+ "grad_norm": 0.010335934348404408,
+ "learning_rate": 0.0016825917315371202,
+ "loss": 4.183531284332275,
+ "step": 1725
+ },
+ {
+ "epoch": 23.978165938864628,
+ "grad_norm": 0.011322776786983013,
+ "learning_rate": 0.001681267247013412,
+ "loss": 4.179635047912598,
+ "step": 1726
+ },
+ {
+ "epoch": 23.992139737991266,
+ "grad_norm": 0.009663194417953491,
+ "learning_rate": 0.001679942628531186,
+ "loss": 4.165654182434082,
+ "step": 1727
+ },
+ {
+ "epoch": 24.0,
+ "grad_norm": 0.010590246878564358,
+ "learning_rate": 0.0016786178771576643,
+ "loss": 4.1539459228515625,
+ "step": 1728
+ },
+ {
+ "epoch": 24.0,
+ "eval_loss": 4.511386394500732,
+ "eval_runtime": 51.7751,
+ "eval_samples_per_second": 47.166,
+ "eval_steps_per_second": 1.487,
+ "step": 1728
+ },
+ {
+ "epoch": 24.01397379912664,
+ "grad_norm": 0.01254354789853096,
+ "learning_rate": 0.0016772929939601728,
+ "loss": 4.104002475738525,
+ "step": 1729
+ },
+ {
+ "epoch": 24.027947598253274,
+ "grad_norm": 0.01300987508147955,
+ "learning_rate": 0.0016759679800061483,
+ "loss": 4.127395153045654,
+ "step": 1730
+ },
+ {
+ "epoch": 24.041921397379912,
+ "grad_norm": 0.013325816951692104,
+ "learning_rate": 0.0016746428363631299,
+ "loss": 4.131229400634766,
+ "step": 1731
+ },
+ {
+ "epoch": 24.05589519650655,
+ "grad_norm": 0.014302870258688927,
+ "learning_rate": 0.001673317564098761,
+ "loss": 4.079358100891113,
+ "step": 1732
+ },
+ {
+ "epoch": 24.069868995633186,
+ "grad_norm": 0.013776068575680256,
+ "learning_rate": 0.0016719921642807905,
+ "loss": 4.132065773010254,
+ "step": 1733
+ },
+ {
+ "epoch": 24.083842794759825,
+ "grad_norm": 0.010788560844957829,
+ "learning_rate": 0.0016706666379770689,
+ "loss": 4.134626388549805,
+ "step": 1734
+ },
+ {
+ "epoch": 24.097816593886463,
+ "grad_norm": 0.009766630828380585,
+ "learning_rate": 0.0016693409862555493,
+ "loss": 4.1645426750183105,
+ "step": 1735
+ },
+ {
+ "epoch": 24.111790393013102,
+ "grad_norm": 0.011760029010474682,
+ "learning_rate": 0.0016680152101842842,
+ "loss": 4.110664367675781,
+ "step": 1736
+ },
+ {
+ "epoch": 24.125764192139737,
+ "grad_norm": 0.011047836393117905,
+ "learning_rate": 0.0016666893108314288,
+ "loss": 4.129222869873047,
+ "step": 1737
+ },
+ {
+ "epoch": 24.139737991266376,
+ "grad_norm": 0.009989139623939991,
+ "learning_rate": 0.0016653632892652344,
+ "loss": 4.076835632324219,
+ "step": 1738
+ },
+ {
+ "epoch": 24.153711790393015,
+ "grad_norm": 0.010033783502876759,
+ "learning_rate": 0.001664037146554054,
+ "loss": 4.11555290222168,
+ "step": 1739
+ },
+ {
+ "epoch": 24.16768558951965,
+ "grad_norm": 0.00996427796781063,
+ "learning_rate": 0.001662710883766337,
+ "loss": 4.131463527679443,
+ "step": 1740
+ },
+ {
+ "epoch": 24.18165938864629,
+ "grad_norm": 0.01044023409485817,
+ "learning_rate": 0.0016613845019706288,
+ "loss": 4.098750114440918,
+ "step": 1741
+ },
+ {
+ "epoch": 24.195633187772927,
+ "grad_norm": 0.010033872909843922,
+ "learning_rate": 0.0016600580022355712,
+ "loss": 4.161447048187256,
+ "step": 1742
+ },
+ {
+ "epoch": 24.209606986899562,
+ "grad_norm": 0.009329567663371563,
+ "learning_rate": 0.0016587313856299012,
+ "loss": 4.146491527557373,
+ "step": 1743
+ },
+ {
+ "epoch": 24.2235807860262,
+ "grad_norm": 0.010061589069664478,
+ "learning_rate": 0.0016574046532224504,
+ "loss": 4.107517719268799,
+ "step": 1744
+ },
+ {
+ "epoch": 24.23755458515284,
+ "grad_norm": 0.009491910226643085,
+ "learning_rate": 0.0016560778060821424,
+ "loss": 4.098226547241211,
+ "step": 1745
+ },
+ {
+ "epoch": 24.251528384279474,
+ "grad_norm": 0.008529345504939556,
+ "learning_rate": 0.0016547508452779943,
+ "loss": 4.1247711181640625,
+ "step": 1746
+ },
+ {
+ "epoch": 24.265502183406113,
+ "grad_norm": 0.00936342403292656,
+ "learning_rate": 0.0016534237718791136,
+ "loss": 4.120573043823242,
+ "step": 1747
+ },
+ {
+ "epoch": 24.27947598253275,
+ "grad_norm": 0.010255641303956509,
+ "learning_rate": 0.0016520965869547008,
+ "loss": 4.089067459106445,
+ "step": 1748
+ },
+ {
+ "epoch": 24.29344978165939,
+ "grad_norm": 0.011638758704066277,
+ "learning_rate": 0.001650769291574044,
+ "loss": 4.074847221374512,
+ "step": 1749
+ },
+ {
+ "epoch": 24.307423580786025,
+ "grad_norm": 0.014707857742905617,
+ "learning_rate": 0.0016494418868065214,
+ "loss": 4.153377532958984,
+ "step": 1750
+ },
+ {
+ "epoch": 24.321397379912664,
+ "grad_norm": 0.01650111749768257,
+ "learning_rate": 0.0016481143737215986,
+ "loss": 4.134062767028809,
+ "step": 1751
+ },
+ {
+ "epoch": 24.335371179039303,
+ "grad_norm": 0.014833372086286545,
+ "learning_rate": 0.00164678675338883,
+ "loss": 4.1344828605651855,
+ "step": 1752
+ },
+ {
+ "epoch": 24.349344978165938,
+ "grad_norm": 0.012257886119186878,
+ "learning_rate": 0.0016454590268778538,
+ "loss": 4.098813533782959,
+ "step": 1753
+ },
+ {
+ "epoch": 24.363318777292577,
+ "grad_norm": 0.013508322648704052,
+ "learning_rate": 0.0016441311952583959,
+ "loss": 4.106297492980957,
+ "step": 1754
+ },
+ {
+ "epoch": 24.377292576419215,
+ "grad_norm": 0.01256900280714035,
+ "learning_rate": 0.001642803259600267,
+ "loss": 4.143631935119629,
+ "step": 1755
+ },
+ {
+ "epoch": 24.39126637554585,
+ "grad_norm": 0.009885406121611595,
+ "learning_rate": 0.00164147522097336,
+ "loss": 4.120885848999023,
+ "step": 1756
+ },
+ {
+ "epoch": 24.40524017467249,
+ "grad_norm": 0.012004674412310123,
+ "learning_rate": 0.0016401470804476518,
+ "loss": 4.089243412017822,
+ "step": 1757
+ },
+ {
+ "epoch": 24.419213973799128,
+ "grad_norm": 0.01224028691649437,
+ "learning_rate": 0.0016388188390932012,
+ "loss": 4.077765464782715,
+ "step": 1758
+ },
+ {
+ "epoch": 24.433187772925763,
+ "grad_norm": 0.012511290609836578,
+ "learning_rate": 0.001637490497980149,
+ "loss": 4.161346435546875,
+ "step": 1759
+ },
+ {
+ "epoch": 24.4471615720524,
+ "grad_norm": 0.010701716877520084,
+ "learning_rate": 0.0016361620581787143,
+ "loss": 4.10698938369751,
+ "step": 1760
+ },
+ {
+ "epoch": 24.46113537117904,
+ "grad_norm": 0.010713113471865654,
+ "learning_rate": 0.001634833520759198,
+ "loss": 4.114283561706543,
+ "step": 1761
+ },
+ {
+ "epoch": 24.475109170305675,
+ "grad_norm": 0.011018907651305199,
+ "learning_rate": 0.001633504886791978,
+ "loss": 4.134284019470215,
+ "step": 1762
+ },
+ {
+ "epoch": 24.489082969432314,
+ "grad_norm": 0.011279584839940071,
+ "learning_rate": 0.0016321761573475117,
+ "loss": 4.1456804275512695,
+ "step": 1763
+ },
+ {
+ "epoch": 24.503056768558952,
+ "grad_norm": 0.010642931796610355,
+ "learning_rate": 0.001630847333496332,
+ "loss": 4.160918235778809,
+ "step": 1764
+ },
+ {
+ "epoch": 24.51703056768559,
+ "grad_norm": 0.011069678701460361,
+ "learning_rate": 0.0016295184163090482,
+ "loss": 4.136838912963867,
+ "step": 1765
+ },
+ {
+ "epoch": 24.531004366812226,
+ "grad_norm": 0.011007166467607021,
+ "learning_rate": 0.0016281894068563443,
+ "loss": 4.129153728485107,
+ "step": 1766
+ },
+ {
+ "epoch": 24.544978165938865,
+ "grad_norm": 0.010055703110992908,
+ "learning_rate": 0.0016268603062089802,
+ "loss": 4.138665676116943,
+ "step": 1767
+ },
+ {
+ "epoch": 24.558951965065503,
+ "grad_norm": 0.009839230217039585,
+ "learning_rate": 0.0016255311154377875,
+ "loss": 4.1366496086120605,
+ "step": 1768
+ },
+ {
+ "epoch": 24.57292576419214,
+ "grad_norm": 0.011191432364284992,
+ "learning_rate": 0.0016242018356136714,
+ "loss": 4.129880905151367,
+ "step": 1769
+ },
+ {
+ "epoch": 24.586899563318777,
+ "grad_norm": 0.012130341492593288,
+ "learning_rate": 0.0016228724678076095,
+ "loss": 4.1425981521606445,
+ "step": 1770
+ },
+ {
+ "epoch": 24.600873362445416,
+ "grad_norm": 0.012707795947790146,
+ "learning_rate": 0.0016215430130906482,
+ "loss": 4.115452766418457,
+ "step": 1771
+ },
+ {
+ "epoch": 24.61484716157205,
+ "grad_norm": 0.012707809917628765,
+ "learning_rate": 0.001620213472533906,
+ "loss": 4.122758865356445,
+ "step": 1772
+ },
+ {
+ "epoch": 24.62882096069869,
+ "grad_norm": 0.01182397548109293,
+ "learning_rate": 0.001618883847208569,
+ "loss": 4.140275955200195,
+ "step": 1773
+ },
+ {
+ "epoch": 24.64279475982533,
+ "grad_norm": 0.010865340940654278,
+ "learning_rate": 0.001617554138185893,
+ "loss": 4.137759208679199,
+ "step": 1774
+ },
+ {
+ "epoch": 24.656768558951963,
+ "grad_norm": 0.010973592288792133,
+ "learning_rate": 0.0016162243465371998,
+ "loss": 4.138947010040283,
+ "step": 1775
+ },
+ {
+ "epoch": 24.670742358078602,
+ "grad_norm": 0.011247226968407631,
+ "learning_rate": 0.001614894473333879,
+ "loss": 4.115516662597656,
+ "step": 1776
+ },
+ {
+ "epoch": 24.68471615720524,
+ "grad_norm": 0.010837487876415253,
+ "learning_rate": 0.0016135645196473853,
+ "loss": 4.160427093505859,
+ "step": 1777
+ },
+ {
+ "epoch": 24.69868995633188,
+ "grad_norm": 0.009096396155655384,
+ "learning_rate": 0.0016122344865492385,
+ "loss": 4.1200032234191895,
+ "step": 1778
+ },
+ {
+ "epoch": 24.712663755458514,
+ "grad_norm": 0.010671247728168964,
+ "learning_rate": 0.0016109043751110223,
+ "loss": 4.097919464111328,
+ "step": 1779
+ },
+ {
+ "epoch": 24.726637554585153,
+ "grad_norm": 0.0107562355697155,
+ "learning_rate": 0.0016095741864043831,
+ "loss": 4.096290588378906,
+ "step": 1780
+ },
+ {
+ "epoch": 24.74061135371179,
+ "grad_norm": 0.008909327909350395,
+ "learning_rate": 0.0016082439215010305,
+ "loss": 4.171017646789551,
+ "step": 1781
+ },
+ {
+ "epoch": 24.754585152838427,
+ "grad_norm": 0.010305124334990978,
+ "learning_rate": 0.0016069135814727343,
+ "loss": 4.1450276374816895,
+ "step": 1782
+ },
+ {
+ "epoch": 24.768558951965066,
+ "grad_norm": 0.011235613375902176,
+ "learning_rate": 0.0016055831673913255,
+ "loss": 4.134704113006592,
+ "step": 1783
+ },
+ {
+ "epoch": 24.782532751091704,
+ "grad_norm": 0.00825632456690073,
+ "learning_rate": 0.0016042526803286949,
+ "loss": 4.107443332672119,
+ "step": 1784
+ },
+ {
+ "epoch": 24.79650655021834,
+ "grad_norm": 0.009056378155946732,
+ "learning_rate": 0.0016029221213567921,
+ "loss": 4.132656097412109,
+ "step": 1785
+ },
+ {
+ "epoch": 24.810480349344978,
+ "grad_norm": 0.008835164830088615,
+ "learning_rate": 0.001601591491547624,
+ "loss": 4.124729633331299,
+ "step": 1786
+ },
+ {
+ "epoch": 24.824454148471617,
+ "grad_norm": 0.00915333442389965,
+ "learning_rate": 0.0016002607919732557,
+ "loss": 4.119500160217285,
+ "step": 1787
+ },
+ {
+ "epoch": 24.83842794759825,
+ "grad_norm": 0.008532539941370487,
+ "learning_rate": 0.0015989300237058066,
+ "loss": 4.141661643981934,
+ "step": 1788
+ },
+ {
+ "epoch": 24.85240174672489,
+ "grad_norm": 0.009952311404049397,
+ "learning_rate": 0.0015975991878174533,
+ "loss": 4.140929222106934,
+ "step": 1789
+ },
+ {
+ "epoch": 24.86637554585153,
+ "grad_norm": 0.008920297026634216,
+ "learning_rate": 0.0015962682853804268,
+ "loss": 4.155877113342285,
+ "step": 1790
+ },
+ {
+ "epoch": 24.880349344978168,
+ "grad_norm": 0.008770049549639225,
+ "learning_rate": 0.0015949373174670104,
+ "loss": 4.126394271850586,
+ "step": 1791
+ },
+ {
+ "epoch": 24.894323144104803,
+ "grad_norm": 0.0078084105625748634,
+ "learning_rate": 0.0015936062851495411,
+ "loss": 4.128876209259033,
+ "step": 1792
+ },
+ {
+ "epoch": 24.90829694323144,
+ "grad_norm": 0.00816641841083765,
+ "learning_rate": 0.0015922751895004078,
+ "loss": 4.137410640716553,
+ "step": 1793
+ },
+ {
+ "epoch": 24.92227074235808,
+ "grad_norm": 0.008895459584891796,
+ "learning_rate": 0.0015909440315920505,
+ "loss": 4.07957649230957,
+ "step": 1794
+ },
+ {
+ "epoch": 24.936244541484715,
+ "grad_norm": 0.008551528677344322,
+ "learning_rate": 0.001589612812496958,
+ "loss": 4.104213714599609,
+ "step": 1795
+ },
+ {
+ "epoch": 24.950218340611354,
+ "grad_norm": 0.008325488306581974,
+ "learning_rate": 0.0015882815332876707,
+ "loss": 4.1584792137146,
+ "step": 1796
+ },
+ {
+ "epoch": 24.964192139737992,
+ "grad_norm": 0.009341313503682613,
+ "learning_rate": 0.0015869501950367744,
+ "loss": 4.1417436599731445,
+ "step": 1797
+ },
+ {
+ "epoch": 24.978165938864628,
+ "grad_norm": 0.00999050959944725,
+ "learning_rate": 0.0015856187988169067,
+ "loss": 4.136049270629883,
+ "step": 1798
+ },
+ {
+ "epoch": 24.992139737991266,
+ "grad_norm": 0.008889932185411453,
+ "learning_rate": 0.0015842873457007484,
+ "loss": 4.16257381439209,
+ "step": 1799
+ },
+ {
+ "epoch": 25.0,
+ "grad_norm": 0.009109573438763618,
+ "learning_rate": 0.0015829558367610268,
+ "loss": 4.171441555023193,
+ "step": 1800
+ },
+ {
+ "epoch": 25.0,
+ "eval_loss": 4.485942840576172,
+ "eval_runtime": 52.1051,
+ "eval_samples_per_second": 46.867,
+ "eval_steps_per_second": 1.478,
+ "step": 1800
+ },
+ {
+ "epoch": 25.01397379912664,
+ "grad_norm": 0.009881424717605114,
+ "learning_rate": 0.001581624273070515,
+ "loss": 4.059608459472656,
+ "step": 1801
+ },
+ {
+ "epoch": 25.027947598253274,
+ "grad_norm": 0.011156721040606499,
+ "learning_rate": 0.0015802926557020294,
+ "loss": 4.047014236450195,
+ "step": 1802
+ },
+ {
+ "epoch": 25.041921397379912,
+ "grad_norm": 0.012326992116868496,
+ "learning_rate": 0.0015789609857284309,
+ "loss": 4.06541109085083,
+ "step": 1803
+ },
+ {
+ "epoch": 25.05589519650655,
+ "grad_norm": 0.01332208700478077,
+ "learning_rate": 0.0015776292642226213,
+ "loss": 4.042714595794678,
+ "step": 1804
+ },
+ {
+ "epoch": 25.069868995633186,
+ "grad_norm": 0.012720128521323204,
+ "learning_rate": 0.0015762974922575446,
+ "loss": 4.098389625549316,
+ "step": 1805
+ },
+ {
+ "epoch": 25.083842794759825,
+ "grad_norm": 0.010723215527832508,
+ "learning_rate": 0.0015749656709061854,
+ "loss": 4.059185981750488,
+ "step": 1806
+ },
+ {
+ "epoch": 25.097816593886463,
+ "grad_norm": 0.010605281218886375,
+ "learning_rate": 0.001573633801241569,
+ "loss": 4.099538326263428,
+ "step": 1807
+ },
+ {
+ "epoch": 25.111790393013102,
+ "grad_norm": 0.01101530622690916,
+ "learning_rate": 0.0015723018843367576,
+ "loss": 4.095674514770508,
+ "step": 1808
+ },
+ {
+ "epoch": 25.125764192139737,
+ "grad_norm": 0.01135838869959116,
+ "learning_rate": 0.0015709699212648532,
+ "loss": 4.070535659790039,
+ "step": 1809
+ },
+ {
+ "epoch": 25.139737991266376,
+ "grad_norm": 0.012119271792471409,
+ "learning_rate": 0.0015696379130989936,
+ "loss": 4.060605049133301,
+ "step": 1810
+ },
+ {
+ "epoch": 25.153711790393015,
+ "grad_norm": 0.012783961370587349,
+ "learning_rate": 0.001568305860912355,
+ "loss": 4.075931072235107,
+ "step": 1811
+ },
+ {
+ "epoch": 25.16768558951965,
+ "grad_norm": 0.011486220173537731,
+ "learning_rate": 0.0015669737657781461,
+ "loss": 4.075855255126953,
+ "step": 1812
+ },
+ {
+ "epoch": 25.18165938864629,
+ "grad_norm": 0.01037552859634161,
+ "learning_rate": 0.0015656416287696136,
+ "loss": 4.087729454040527,
+ "step": 1813
+ },
+ {
+ "epoch": 25.195633187772927,
+ "grad_norm": 0.009802743792533875,
+ "learning_rate": 0.0015643094509600354,
+ "loss": 4.0855607986450195,
+ "step": 1814
+ },
+ {
+ "epoch": 25.209606986899562,
+ "grad_norm": 0.010804438032209873,
+ "learning_rate": 0.001562977233422723,
+ "loss": 4.064528465270996,
+ "step": 1815
+ },
+ {
+ "epoch": 25.2235807860262,
+ "grad_norm": 0.009505728259682655,
+ "learning_rate": 0.0015616449772310206,
+ "loss": 4.036775588989258,
+ "step": 1816
+ },
+ {
+ "epoch": 25.23755458515284,
+ "grad_norm": 0.010634485632181168,
+ "learning_rate": 0.0015603126834583024,
+ "loss": 4.069923400878906,
+ "step": 1817
+ },
+ {
+ "epoch": 25.251528384279474,
+ "grad_norm": 0.011392096057534218,
+ "learning_rate": 0.0015589803531779736,
+ "loss": 4.0812177658081055,
+ "step": 1818
+ },
+ {
+ "epoch": 25.265502183406113,
+ "grad_norm": 0.011851722374558449,
+ "learning_rate": 0.0015576479874634683,
+ "loss": 4.09183406829834,
+ "step": 1819
+ },
+ {
+ "epoch": 25.27947598253275,
+ "grad_norm": 0.011205444112420082,
+ "learning_rate": 0.0015563155873882502,
+ "loss": 4.07825231552124,
+ "step": 1820
+ },
+ {
+ "epoch": 25.29344978165939,
+ "grad_norm": 0.009676349349319935,
+ "learning_rate": 0.0015549831540258093,
+ "loss": 4.052826404571533,
+ "step": 1821
+ },
+ {
+ "epoch": 25.307423580786025,
+ "grad_norm": 0.009376165457069874,
+ "learning_rate": 0.0015536506884496638,
+ "loss": 4.02100133895874,
+ "step": 1822
+ },
+ {
+ "epoch": 25.321397379912664,
+ "grad_norm": 0.00936955213546753,
+ "learning_rate": 0.0015523181917333564,
+ "loss": 4.049526214599609,
+ "step": 1823
+ },
+ {
+ "epoch": 25.335371179039303,
+ "grad_norm": 0.00937784556299448,
+ "learning_rate": 0.0015509856649504564,
+ "loss": 4.096356391906738,
+ "step": 1824
+ },
+ {
+ "epoch": 25.349344978165938,
+ "grad_norm": 0.010014008730649948,
+ "learning_rate": 0.0015496531091745555,
+ "loss": 4.03254508972168,
+ "step": 1825
+ },
+ {
+ "epoch": 25.363318777292577,
+ "grad_norm": 0.009909634478390217,
+ "learning_rate": 0.0015483205254792705,
+ "loss": 4.067630767822266,
+ "step": 1826
+ },
+ {
+ "epoch": 25.377292576419215,
+ "grad_norm": 0.009319501928985119,
+ "learning_rate": 0.0015469879149382399,
+ "loss": 4.094167709350586,
+ "step": 1827
+ },
+ {
+ "epoch": 25.39126637554585,
+ "grad_norm": 0.008634721860289574,
+ "learning_rate": 0.0015456552786251244,
+ "loss": 4.122404098510742,
+ "step": 1828
+ },
+ {
+ "epoch": 25.40524017467249,
+ "grad_norm": 0.009412454441189766,
+ "learning_rate": 0.0015443226176136049,
+ "loss": 4.081546783447266,
+ "step": 1829
+ },
+ {
+ "epoch": 25.419213973799128,
+ "grad_norm": 0.010102108120918274,
+ "learning_rate": 0.0015429899329773814,
+ "loss": 4.049839019775391,
+ "step": 1830
+ },
+ {
+ "epoch": 25.433187772925763,
+ "grad_norm": 0.010921654291450977,
+ "learning_rate": 0.0015416572257901746,
+ "loss": 4.097097396850586,
+ "step": 1831
+ },
+ {
+ "epoch": 25.4471615720524,
+ "grad_norm": 0.010607713833451271,
+ "learning_rate": 0.0015403244971257226,
+ "loss": 4.084481239318848,
+ "step": 1832
+ },
+ {
+ "epoch": 25.46113537117904,
+ "grad_norm": 0.011887960135936737,
+ "learning_rate": 0.0015389917480577806,
+ "loss": 4.074968338012695,
+ "step": 1833
+ },
+ {
+ "epoch": 25.475109170305675,
+ "grad_norm": 0.01089888159185648,
+ "learning_rate": 0.0015376589796601203,
+ "loss": 4.062132835388184,
+ "step": 1834
+ },
+ {
+ "epoch": 25.489082969432314,
+ "grad_norm": 0.009934183210134506,
+ "learning_rate": 0.0015363261930065298,
+ "loss": 4.128092288970947,
+ "step": 1835
+ },
+ {
+ "epoch": 25.503056768558952,
+ "grad_norm": 0.00878534372895956,
+ "learning_rate": 0.0015349933891708105,
+ "loss": 4.069051742553711,
+ "step": 1836
+ },
+ {
+ "epoch": 25.51703056768559,
+ "grad_norm": 0.009928155690431595,
+ "learning_rate": 0.0015336605692267788,
+ "loss": 4.073912620544434,
+ "step": 1837
+ },
+ {
+ "epoch": 25.531004366812226,
+ "grad_norm": 0.011091411113739014,
+ "learning_rate": 0.0015323277342482637,
+ "loss": 4.125944137573242,
+ "step": 1838
+ },
+ {
+ "epoch": 25.544978165938865,
+ "grad_norm": 0.010314172133803368,
+ "learning_rate": 0.0015309948853091055,
+ "loss": 4.062620162963867,
+ "step": 1839
+ },
+ {
+ "epoch": 25.558951965065503,
+ "grad_norm": 0.010423967614769936,
+ "learning_rate": 0.001529662023483158,
+ "loss": 4.101737976074219,
+ "step": 1840
+ },
+ {
+ "epoch": 25.57292576419214,
+ "grad_norm": 0.009822997264564037,
+ "learning_rate": 0.0015283291498442817,
+ "loss": 4.064520359039307,
+ "step": 1841
+ },
+ {
+ "epoch": 25.586899563318777,
+ "grad_norm": 0.009757095016539097,
+ "learning_rate": 0.001526996265466351,
+ "loss": 4.100624084472656,
+ "step": 1842
+ },
+ {
+ "epoch": 25.600873362445416,
+ "grad_norm": 0.008650688454508781,
+ "learning_rate": 0.001525663371423246,
+ "loss": 4.109685897827148,
+ "step": 1843
+ },
+ {
+ "epoch": 25.61484716157205,
+ "grad_norm": 0.007378234062343836,
+ "learning_rate": 0.0015243304687888553,
+ "loss": 4.044726848602295,
+ "step": 1844
+ },
+ {
+ "epoch": 25.62882096069869,
+ "grad_norm": 0.008194336667656898,
+ "learning_rate": 0.001522997558637074,
+ "loss": 4.097923278808594,
+ "step": 1845
+ },
+ {
+ "epoch": 25.64279475982533,
+ "grad_norm": 0.009764892049133778,
+ "learning_rate": 0.0015216646420418048,
+ "loss": 4.102148056030273,
+ "step": 1846
+ },
+ {
+ "epoch": 25.656768558951963,
+ "grad_norm": 0.009242868050932884,
+ "learning_rate": 0.001520331720076953,
+ "loss": 4.068563461303711,
+ "step": 1847
+ },
+ {
+ "epoch": 25.670742358078602,
+ "grad_norm": 0.010409324429929256,
+ "learning_rate": 0.001518998793816431,
+ "loss": 4.065944194793701,
+ "step": 1848
+ },
+ {
+ "epoch": 25.68471615720524,
+ "grad_norm": 0.009215258061885834,
+ "learning_rate": 0.0015176658643341527,
+ "loss": 4.10223388671875,
+ "step": 1849
+ },
+ {
+ "epoch": 25.69868995633188,
+ "grad_norm": 0.008065393194556236,
+ "learning_rate": 0.0015163329327040366,
+ "loss": 4.099432945251465,
+ "step": 1850
+ },
+ {
+ "epoch": 25.712663755458514,
+ "grad_norm": 0.008498288691043854,
+ "learning_rate": 0.001515,
+ "loss": 4.109854698181152,
+ "step": 1851
+ },
+ {
+ "epoch": 25.726637554585153,
+ "grad_norm": 0.008919150568544865,
+ "learning_rate": 0.0015136670672959633,
+ "loss": 4.128247261047363,
+ "step": 1852
+ },
+ {
+ "epoch": 25.74061135371179,
+ "grad_norm": 0.009104429744184017,
+ "learning_rate": 0.0015123341356658474,
+ "loss": 4.068622589111328,
+ "step": 1853
+ },
+ {
+ "epoch": 25.754585152838427,
+ "grad_norm": 0.008558365516364574,
+ "learning_rate": 0.0015110012061835692,
+ "loss": 4.114214897155762,
+ "step": 1854
+ },
+ {
+ "epoch": 25.768558951965066,
+ "grad_norm": 0.007865076884627342,
+ "learning_rate": 0.0015096682799230477,
+ "loss": 4.11155366897583,
+ "step": 1855
+ },
+ {
+ "epoch": 25.782532751091704,
+ "grad_norm": 0.007848534733057022,
+ "learning_rate": 0.0015083353579581958,
+ "loss": 4.045694351196289,
+ "step": 1856
+ },
+ {
+ "epoch": 25.79650655021834,
+ "grad_norm": 0.008238030597567558,
+ "learning_rate": 0.0015070024413629264,
+ "loss": 4.093930721282959,
+ "step": 1857
+ },
+ {
+ "epoch": 25.810480349344978,
+ "grad_norm": 0.007536398712545633,
+ "learning_rate": 0.0015056695312111448,
+ "loss": 4.106235504150391,
+ "step": 1858
+ },
+ {
+ "epoch": 25.824454148471617,
+ "grad_norm": 0.008156892843544483,
+ "learning_rate": 0.001504336628576754,
+ "loss": 4.13060998916626,
+ "step": 1859
+ },
+ {
+ "epoch": 25.83842794759825,
+ "grad_norm": 0.008144439198076725,
+ "learning_rate": 0.0015030037345336487,
+ "loss": 4.091304779052734,
+ "step": 1860
+ },
+ {
+ "epoch": 25.85240174672489,
+ "grad_norm": 0.007757358253002167,
+ "learning_rate": 0.001501670850155718,
+ "loss": 4.069276809692383,
+ "step": 1861
+ },
+ {
+ "epoch": 25.86637554585153,
+ "grad_norm": 0.00860623549669981,
+ "learning_rate": 0.0015003379765168425,
+ "loss": 4.107658386230469,
+ "step": 1862
+ },
+ {
+ "epoch": 25.880349344978168,
+ "grad_norm": 0.008614457212388515,
+ "learning_rate": 0.0014990051146908949,
+ "loss": 4.106705665588379,
+ "step": 1863
+ },
+ {
+ "epoch": 25.894323144104803,
+ "grad_norm": 0.009120592847466469,
+ "learning_rate": 0.001497672265751737,
+ "loss": 4.0891923904418945,
+ "step": 1864
+ },
+ {
+ "epoch": 25.90829694323144,
+ "grad_norm": 0.009446980431675911,
+ "learning_rate": 0.0014963394307732213,
+ "loss": 4.098847389221191,
+ "step": 1865
+ },
+ {
+ "epoch": 25.92227074235808,
+ "grad_norm": 0.010494766756892204,
+ "learning_rate": 0.0014950066108291894,
+ "loss": 4.115325450897217,
+ "step": 1866
+ },
+ {
+ "epoch": 25.936244541484715,
+ "grad_norm": 0.010057658888399601,
+ "learning_rate": 0.0014936738069934706,
+ "loss": 4.099984169006348,
+ "step": 1867
+ },
+ {
+ "epoch": 25.950218340611354,
+ "grad_norm": 0.009071771055459976,
+ "learning_rate": 0.0014923410203398796,
+ "loss": 4.091543674468994,
+ "step": 1868
+ },
+ {
+ "epoch": 25.964192139737992,
+ "grad_norm": 0.009937075898051262,
+ "learning_rate": 0.00149100825194222,
+ "loss": 4.086492538452148,
+ "step": 1869
+ },
+ {
+ "epoch": 25.978165938864628,
+ "grad_norm": 0.010222986340522766,
+ "learning_rate": 0.0014896755028742778,
+ "loss": 4.063841819763184,
+ "step": 1870
+ },
+ {
+ "epoch": 25.992139737991266,
+ "grad_norm": 0.010513762012124062,
+ "learning_rate": 0.001488342774209826,
+ "loss": 4.11991548538208,
+ "step": 1871
+ },
+ {
+ "epoch": 26.0,
+ "grad_norm": 0.009530222974717617,
+ "learning_rate": 0.0014870100670226192,
+ "loss": 4.043334007263184,
+ "step": 1872
+ },
+ {
+ "epoch": 26.0,
+ "eval_loss": 4.467459678649902,
+ "eval_runtime": 54.1628,
+ "eval_samples_per_second": 45.086,
+ "eval_steps_per_second": 1.422,
+ "step": 1872
+ },
+ {
+ "epoch": 26.01397379912664,
+ "grad_norm": 0.010274187661707401,
+ "learning_rate": 0.0014856773823863955,
+ "loss": 4.049561500549316,
+ "step": 1873
+ },
+ {
+ "epoch": 26.027947598253274,
+ "grad_norm": 0.011448442004621029,
+ "learning_rate": 0.001484344721374876,
+ "loss": 4.019484996795654,
+ "step": 1874
+ },
+ {
+ "epoch": 26.041921397379912,
+ "grad_norm": 0.011361192911863327,
+ "learning_rate": 0.0014830120850617599,
+ "loss": 4.0429887771606445,
+ "step": 1875
+ },
+ {
+ "epoch": 26.05589519650655,
+ "grad_norm": 0.010250156745314598,
+ "learning_rate": 0.0014816794745207299,
+ "loss": 4.012682914733887,
+ "step": 1876
+ },
+ {
+ "epoch": 26.069868995633186,
+ "grad_norm": 0.010024400427937508,
+ "learning_rate": 0.0014803468908254449,
+ "loss": 3.979409694671631,
+ "step": 1877
+ },
+ {
+ "epoch": 26.083842794759825,
+ "grad_norm": 0.01153535582125187,
+ "learning_rate": 0.0014790143350495446,
+ "loss": 4.045780181884766,
+ "step": 1878
+ },
+ {
+ "epoch": 26.097816593886463,
+ "grad_norm": 0.012632817961275578,
+ "learning_rate": 0.001477681808266644,
+ "loss": 4.057641983032227,
+ "step": 1879
+ },
+ {
+ "epoch": 26.111790393013102,
+ "grad_norm": 0.013758293353021145,
+ "learning_rate": 0.0014763493115503366,
+ "loss": 4.061607360839844,
+ "step": 1880
+ },
+ {
+ "epoch": 26.125764192139737,
+ "grad_norm": 0.014041583053767681,
+ "learning_rate": 0.001475016845974191,
+ "loss": 4.035029411315918,
+ "step": 1881
+ },
+ {
+ "epoch": 26.139737991266376,
+ "grad_norm": 0.013629456982016563,
+ "learning_rate": 0.0014736844126117497,
+ "loss": 4.062169075012207,
+ "step": 1882
+ },
+ {
+ "epoch": 26.153711790393015,
+ "grad_norm": 0.011023213155567646,
+ "learning_rate": 0.0014723520125365323,
+ "loss": 4.02731990814209,
+ "step": 1883
+ },
+ {
+ "epoch": 26.16768558951965,
+ "grad_norm": 0.01001750584691763,
+ "learning_rate": 0.0014710196468220266,
+ "loss": 4.019882678985596,
+ "step": 1884
+ },
+ {
+ "epoch": 26.18165938864629,
+ "grad_norm": 0.008321646600961685,
+ "learning_rate": 0.001469687316541698,
+ "loss": 4.005446434020996,
+ "step": 1885
+ },
+ {
+ "epoch": 26.195633187772927,
+ "grad_norm": 0.008935946971178055,
+ "learning_rate": 0.0014683550227689796,
+ "loss": 4.017102241516113,
+ "step": 1886
+ },
+ {
+ "epoch": 26.209606986899562,
+ "grad_norm": 0.009137130342423916,
+ "learning_rate": 0.001467022766577277,
+ "loss": 4.024102210998535,
+ "step": 1887
+ },
+ {
+ "epoch": 26.2235807860262,
+ "grad_norm": 0.00805022194981575,
+ "learning_rate": 0.001465690549039965,
+ "loss": 3.9935250282287598,
+ "step": 1888
+ },
+ {
+ "epoch": 26.23755458515284,
+ "grad_norm": 0.009204058907926083,
+ "learning_rate": 0.0014643583712303865,
+ "loss": 4.042818546295166,
+ "step": 1889
+ },
+ {
+ "epoch": 26.251528384279474,
+ "grad_norm": 0.009812815114855766,
+ "learning_rate": 0.001463026234221854,
+ "loss": 4.0252485275268555,
+ "step": 1890
+ },
+ {
+ "epoch": 26.265502183406113,
+ "grad_norm": 0.009955782443284988,
+ "learning_rate": 0.0014616941390876454,
+ "loss": 3.9839768409729004,
+ "step": 1891
+ },
+ {
+ "epoch": 26.27947598253275,
+ "grad_norm": 0.009374645538628101,
+ "learning_rate": 0.001460362086901007,
+ "loss": 4.040245056152344,
+ "step": 1892
+ },
+ {
+ "epoch": 26.29344978165939,
+ "grad_norm": 0.010008739307522774,
+ "learning_rate": 0.0014590300787351471,
+ "loss": 4.0306596755981445,
+ "step": 1893
+ },
+ {
+ "epoch": 26.307423580786025,
+ "grad_norm": 0.009014200419187546,
+ "learning_rate": 0.001457698115663243,
+ "loss": 4.028713226318359,
+ "step": 1894
+ },
+ {
+ "epoch": 26.321397379912664,
+ "grad_norm": 0.008381523191928864,
+ "learning_rate": 0.0014563661987584313,
+ "loss": 4.024718284606934,
+ "step": 1895
+ },
+ {
+ "epoch": 26.335371179039303,
+ "grad_norm": 0.008627948351204395,
+ "learning_rate": 0.0014550343290938143,
+ "loss": 4.02303409576416,
+ "step": 1896
+ },
+ {
+ "epoch": 26.349344978165938,
+ "grad_norm": 0.010257705114781857,
+ "learning_rate": 0.0014537025077424556,
+ "loss": 4.061058044433594,
+ "step": 1897
+ },
+ {
+ "epoch": 26.363318777292577,
+ "grad_norm": 0.010759209282696247,
+ "learning_rate": 0.0014523707357773788,
+ "loss": 4.064718246459961,
+ "step": 1898
+ },
+ {
+ "epoch": 26.377292576419215,
+ "grad_norm": 0.009966139681637287,
+ "learning_rate": 0.0014510390142715697,
+ "loss": 3.9919447898864746,
+ "step": 1899
+ },
+ {
+ "epoch": 26.39126637554585,
+ "grad_norm": 0.00997584406286478,
+ "learning_rate": 0.001449707344297971,
+ "loss": 4.0431599617004395,
+ "step": 1900
+ },
+ {
+ "epoch": 26.40524017467249,
+ "grad_norm": 0.010846944525837898,
+ "learning_rate": 0.0014483757269294858,
+ "loss": 4.0222697257995605,
+ "step": 1901
+ },
+ {
+ "epoch": 26.419213973799128,
+ "grad_norm": 0.00991352554410696,
+ "learning_rate": 0.0014470441632389735,
+ "loss": 4.059438705444336,
+ "step": 1902
+ },
+ {
+ "epoch": 26.433187772925763,
+ "grad_norm": 0.008994397707283497,
+ "learning_rate": 0.0014457126542992518,
+ "loss": 4.021494388580322,
+ "step": 1903
+ },
+ {
+ "epoch": 26.4471615720524,
+ "grad_norm": 0.009567267261445522,
+ "learning_rate": 0.0014443812011830934,
+ "loss": 4.070810317993164,
+ "step": 1904
+ },
+ {
+ "epoch": 26.46113537117904,
+ "grad_norm": 0.009199011139571667,
+ "learning_rate": 0.0014430498049632255,
+ "loss": 4.020938873291016,
+ "step": 1905
+ },
+ {
+ "epoch": 26.475109170305675,
+ "grad_norm": 0.010440350510179996,
+ "learning_rate": 0.0014417184667123303,
+ "loss": 4.062529563903809,
+ "step": 1906
+ },
+ {
+ "epoch": 26.489082969432314,
+ "grad_norm": 0.009962482377886772,
+ "learning_rate": 0.0014403871875030424,
+ "loss": 4.0687150955200195,
+ "step": 1907
+ },
+ {
+ "epoch": 26.503056768558952,
+ "grad_norm": 0.009389379993081093,
+ "learning_rate": 0.0014390559684079502,
+ "loss": 4.055185794830322,
+ "step": 1908
+ },
+ {
+ "epoch": 26.51703056768559,
+ "grad_norm": 0.011469941586256027,
+ "learning_rate": 0.0014377248104995924,
+ "loss": 4.041023254394531,
+ "step": 1909
+ },
+ {
+ "epoch": 26.531004366812226,
+ "grad_norm": 0.011034650728106499,
+ "learning_rate": 0.0014363937148504588,
+ "loss": 4.054254531860352,
+ "step": 1910
+ },
+ {
+ "epoch": 26.544978165938865,
+ "grad_norm": 0.010651151649653912,
+ "learning_rate": 0.0014350626825329897,
+ "loss": 4.037753582000732,
+ "step": 1911
+ },
+ {
+ "epoch": 26.558951965065503,
+ "grad_norm": 0.010747930034995079,
+ "learning_rate": 0.0014337317146195733,
+ "loss": 4.059938907623291,
+ "step": 1912
+ },
+ {
+ "epoch": 26.57292576419214,
+ "grad_norm": 0.00987576600164175,
+ "learning_rate": 0.0014324008121825468,
+ "loss": 4.075366973876953,
+ "step": 1913
+ },
+ {
+ "epoch": 26.586899563318777,
+ "grad_norm": 0.008504371158778667,
+ "learning_rate": 0.0014310699762941936,
+ "loss": 4.026755332946777,
+ "step": 1914
+ },
+ {
+ "epoch": 26.600873362445416,
+ "grad_norm": 0.009445052593946457,
+ "learning_rate": 0.001429739208026745,
+ "loss": 4.034431457519531,
+ "step": 1915
+ },
+ {
+ "epoch": 26.61484716157205,
+ "grad_norm": 0.010560023598372936,
+ "learning_rate": 0.0014284085084523763,
+ "loss": 4.101158618927002,
+ "step": 1916
+ },
+ {
+ "epoch": 26.62882096069869,
+ "grad_norm": 0.009920580312609673,
+ "learning_rate": 0.0014270778786432078,
+ "loss": 4.041321754455566,
+ "step": 1917
+ },
+ {
+ "epoch": 26.64279475982533,
+ "grad_norm": 0.009854268282651901,
+ "learning_rate": 0.0014257473196713053,
+ "loss": 4.049882411956787,
+ "step": 1918
+ },
+ {
+ "epoch": 26.656768558951963,
+ "grad_norm": 0.009825135581195354,
+ "learning_rate": 0.0014244168326086745,
+ "loss": 4.067440986633301,
+ "step": 1919
+ },
+ {
+ "epoch": 26.670742358078602,
+ "grad_norm": 0.010506784543395042,
+ "learning_rate": 0.0014230864185272663,
+ "loss": 4.042476654052734,
+ "step": 1920
+ },
+ {
+ "epoch": 26.68471615720524,
+ "grad_norm": 0.009527767077088356,
+ "learning_rate": 0.0014217560784989696,
+ "loss": 4.055248260498047,
+ "step": 1921
+ },
+ {
+ "epoch": 26.69868995633188,
+ "grad_norm": 0.01013826485723257,
+ "learning_rate": 0.001420425813595617,
+ "loss": 4.05246639251709,
+ "step": 1922
+ },
+ {
+ "epoch": 26.712663755458514,
+ "grad_norm": 0.011020111851394176,
+ "learning_rate": 0.0014190956248889778,
+ "loss": 4.032987117767334,
+ "step": 1923
+ },
+ {
+ "epoch": 26.726637554585153,
+ "grad_norm": 0.010647117160260677,
+ "learning_rate": 0.0014177655134507614,
+ "loss": 4.093589782714844,
+ "step": 1924
+ },
+ {
+ "epoch": 26.74061135371179,
+ "grad_norm": 0.009484780952334404,
+ "learning_rate": 0.001416435480352615,
+ "loss": 4.017437934875488,
+ "step": 1925
+ },
+ {
+ "epoch": 26.754585152838427,
+ "grad_norm": 0.0114199910312891,
+ "learning_rate": 0.0014151055266661211,
+ "loss": 4.080643653869629,
+ "step": 1926
+ },
+ {
+ "epoch": 26.768558951965066,
+ "grad_norm": 0.011137664318084717,
+ "learning_rate": 0.0014137756534628007,
+ "loss": 4.069380760192871,
+ "step": 1927
+ },
+ {
+ "epoch": 26.782532751091704,
+ "grad_norm": 0.008972682990133762,
+ "learning_rate": 0.0014124458618141073,
+ "loss": 4.051383018493652,
+ "step": 1928
+ },
+ {
+ "epoch": 26.79650655021834,
+ "grad_norm": 0.010311643593013287,
+ "learning_rate": 0.0014111161527914317,
+ "loss": 4.048953533172607,
+ "step": 1929
+ },
+ {
+ "epoch": 26.810480349344978,
+ "grad_norm": 0.01113453321158886,
+ "learning_rate": 0.0014097865274660944,
+ "loss": 4.063544273376465,
+ "step": 1930
+ },
+ {
+ "epoch": 26.824454148471617,
+ "grad_norm": 0.01091703586280346,
+ "learning_rate": 0.0014084569869093518,
+ "loss": 4.006721496582031,
+ "step": 1931
+ },
+ {
+ "epoch": 26.83842794759825,
+ "grad_norm": 0.010672449134290218,
+ "learning_rate": 0.0014071275321923906,
+ "loss": 4.020220756530762,
+ "step": 1932
+ },
+ {
+ "epoch": 26.85240174672489,
+ "grad_norm": 0.009682527743279934,
+ "learning_rate": 0.001405798164386328,
+ "loss": 4.057307243347168,
+ "step": 1933
+ },
+ {
+ "epoch": 26.86637554585153,
+ "grad_norm": 0.009808676317334175,
+ "learning_rate": 0.0014044688845622126,
+ "loss": 4.105412483215332,
+ "step": 1934
+ },
+ {
+ "epoch": 26.880349344978168,
+ "grad_norm": 0.010351189412176609,
+ "learning_rate": 0.0014031396937910202,
+ "loss": 4.066454887390137,
+ "step": 1935
+ },
+ {
+ "epoch": 26.894323144104803,
+ "grad_norm": 0.010038275271654129,
+ "learning_rate": 0.001401810593143656,
+ "loss": 4.056752681732178,
+ "step": 1936
+ },
+ {
+ "epoch": 26.90829694323144,
+ "grad_norm": 0.01016094908118248,
+ "learning_rate": 0.001400481583690952,
+ "loss": 4.097933292388916,
+ "step": 1937
+ },
+ {
+ "epoch": 26.92227074235808,
+ "grad_norm": 0.008683547377586365,
+ "learning_rate": 0.0013991526665036679,
+ "loss": 4.0386152267456055,
+ "step": 1938
+ },
+ {
+ "epoch": 26.936244541484715,
+ "grad_norm": 0.009556332603096962,
+ "learning_rate": 0.0013978238426524882,
+ "loss": 4.05835485458374,
+ "step": 1939
+ },
+ {
+ "epoch": 26.950218340611354,
+ "grad_norm": 0.009597044438123703,
+ "learning_rate": 0.001396495113208022,
+ "loss": 4.052363872528076,
+ "step": 1940
+ },
+ {
+ "epoch": 26.964192139737992,
+ "grad_norm": 0.010021139867603779,
+ "learning_rate": 0.0013951664792408026,
+ "loss": 4.083431243896484,
+ "step": 1941
+ },
+ {
+ "epoch": 26.978165938864628,
+ "grad_norm": 0.010611358098685741,
+ "learning_rate": 0.0013938379418212863,
+ "loss": 4.042330265045166,
+ "step": 1942
+ },
+ {
+ "epoch": 26.992139737991266,
+ "grad_norm": 0.009132112376391888,
+ "learning_rate": 0.001392509502019852,
+ "loss": 4.058825969696045,
+ "step": 1943
+ },
+ {
+ "epoch": 27.0,
+ "grad_norm": 0.009808709844946861,
+ "learning_rate": 0.0013911811609067992,
+ "loss": 4.093010902404785,
+ "step": 1944
+ },
+ {
+ "epoch": 27.0,
+ "eval_loss": 4.4473161697387695,
+ "eval_runtime": 51.7738,
+ "eval_samples_per_second": 47.167,
+ "eval_steps_per_second": 1.487,
+ "step": 1944
+ },
+ {
+ "epoch": 27.01397379912664,
+ "grad_norm": 0.011859052814543247,
+ "learning_rate": 0.0013898529195523488,
+ "loss": 3.9572651386260986,
+ "step": 1945
+ },
+ {
+ "epoch": 27.027947598253274,
+ "grad_norm": 0.013210238888859749,
+ "learning_rate": 0.0013885247790266406,
+ "loss": 3.9812350273132324,
+ "step": 1946
+ },
+ {
+ "epoch": 27.041921397379912,
+ "grad_norm": 0.012981324456632137,
+ "learning_rate": 0.001387196740399733,
+ "loss": 3.987725019454956,
+ "step": 1947
+ },
+ {
+ "epoch": 27.05589519650655,
+ "grad_norm": 0.01424382347613573,
+ "learning_rate": 0.001385868804741604,
+ "loss": 3.9719889163970947,
+ "step": 1948
+ },
+ {
+ "epoch": 27.069868995633186,
+ "grad_norm": 0.0154642965644598,
+ "learning_rate": 0.0013845409731221465,
+ "loss": 3.984248161315918,
+ "step": 1949
+ },
+ {
+ "epoch": 27.083842794759825,
+ "grad_norm": 0.014779271557927132,
+ "learning_rate": 0.0013832132466111706,
+ "loss": 4.002912521362305,
+ "step": 1950
+ },
+ {
+ "epoch": 27.097816593886463,
+ "grad_norm": 0.012921188026666641,
+ "learning_rate": 0.0013818856262784013,
+ "loss": 3.9588499069213867,
+ "step": 1951
+ },
+ {
+ "epoch": 27.111790393013102,
+ "grad_norm": 0.01165325753390789,
+ "learning_rate": 0.0013805581131934792,
+ "loss": 3.993288993835449,
+ "step": 1952
+ },
+ {
+ "epoch": 27.125764192139737,
+ "grad_norm": 0.008908954448997974,
+ "learning_rate": 0.0013792307084259563,
+ "loss": 3.987151861190796,
+ "step": 1953
+ },
+ {
+ "epoch": 27.139737991266376,
+ "grad_norm": 0.012040507979691029,
+ "learning_rate": 0.0013779034130452991,
+ "loss": 3.9675703048706055,
+ "step": 1954
+ },
+ {
+ "epoch": 27.153711790393015,
+ "grad_norm": 0.013464349322021008,
+ "learning_rate": 0.0013765762281208868,
+ "loss": 3.9901437759399414,
+ "step": 1955
+ },
+ {
+ "epoch": 27.16768558951965,
+ "grad_norm": 0.009626250714063644,
+ "learning_rate": 0.0013752491547220063,
+ "loss": 3.9373281002044678,
+ "step": 1956
+ },
+ {
+ "epoch": 27.18165938864629,
+ "grad_norm": 0.00897257961332798,
+ "learning_rate": 0.0013739221939178582,
+ "loss": 4.00804328918457,
+ "step": 1957
+ },
+ {
+ "epoch": 27.195633187772927,
+ "grad_norm": 0.010689414106309414,
+ "learning_rate": 0.00137259534677755,
+ "loss": 4.020804405212402,
+ "step": 1958
+ },
+ {
+ "epoch": 27.209606986899562,
+ "grad_norm": 0.010462156496942043,
+ "learning_rate": 0.0013712686143700988,
+ "loss": 4.028973579406738,
+ "step": 1959
+ },
+ {
+ "epoch": 27.2235807860262,
+ "grad_norm": 0.009135857224464417,
+ "learning_rate": 0.001369941997764429,
+ "loss": 3.9798216819763184,
+ "step": 1960
+ },
+ {
+ "epoch": 27.23755458515284,
+ "grad_norm": 0.009298899210989475,
+ "learning_rate": 0.0013686154980293713,
+ "loss": 4.018845081329346,
+ "step": 1961
+ },
+ {
+ "epoch": 27.251528384279474,
+ "grad_norm": 0.00979928020387888,
+ "learning_rate": 0.0013672891162336636,
+ "loss": 3.9946775436401367,
+ "step": 1962
+ },
+ {
+ "epoch": 27.265502183406113,
+ "grad_norm": 0.008175542578101158,
+ "learning_rate": 0.0013659628534459457,
+ "loss": 3.981858491897583,
+ "step": 1963
+ },
+ {
+ "epoch": 27.27947598253275,
+ "grad_norm": 0.008302642032504082,
+ "learning_rate": 0.0013646367107347662,
+ "loss": 3.991128444671631,
+ "step": 1964
+ },
+ {
+ "epoch": 27.29344978165939,
+ "grad_norm": 0.008516009896993637,
+ "learning_rate": 0.0013633106891685718,
+ "loss": 4.004990100860596,
+ "step": 1965
+ },
+ {
+ "epoch": 27.307423580786025,
+ "grad_norm": 0.007936482317745686,
+ "learning_rate": 0.0013619847898157164,
+ "loss": 3.995734214782715,
+ "step": 1966
+ },
+ {
+ "epoch": 27.321397379912664,
+ "grad_norm": 0.008317600004374981,
+ "learning_rate": 0.001360659013744451,
+ "loss": 4.040653228759766,
+ "step": 1967
+ },
+ {
+ "epoch": 27.335371179039303,
+ "grad_norm": 0.008481282740831375,
+ "learning_rate": 0.001359333362022931,
+ "loss": 3.969860553741455,
+ "step": 1968
+ },
+ {
+ "epoch": 27.349344978165938,
+ "grad_norm": 0.008944948203861713,
+ "learning_rate": 0.0013580078357192096,
+ "loss": 3.9908909797668457,
+ "step": 1969
+ },
+ {
+ "epoch": 27.363318777292577,
+ "grad_norm": 0.008897850289940834,
+ "learning_rate": 0.001356682435901239,
+ "loss": 4.04786491394043,
+ "step": 1970
+ },
+ {
+ "epoch": 27.377292576419215,
+ "grad_norm": 0.009527456946671009,
+ "learning_rate": 0.001355357163636871,
+ "loss": 4.017423629760742,
+ "step": 1971
+ },
+ {
+ "epoch": 27.39126637554585,
+ "grad_norm": 0.011714582331478596,
+ "learning_rate": 0.001354032019993852,
+ "loss": 3.9990692138671875,
+ "step": 1972
+ },
+ {
+ "epoch": 27.40524017467249,
+ "grad_norm": 0.011083807796239853,
+ "learning_rate": 0.0013527070060398274,
+ "loss": 3.995265483856201,
+ "step": 1973
+ },
+ {
+ "epoch": 27.419213973799128,
+ "grad_norm": 0.009878288954496384,
+ "learning_rate": 0.0013513821228423363,
+ "loss": 4.0364532470703125,
+ "step": 1974
+ },
+ {
+ "epoch": 27.433187772925763,
+ "grad_norm": 0.009727329947054386,
+ "learning_rate": 0.0013500573714688136,
+ "loss": 4.035953521728516,
+ "step": 1975
+ },
+ {
+ "epoch": 27.4471615720524,
+ "grad_norm": 0.009482033550739288,
+ "learning_rate": 0.0013487327529865879,
+ "loss": 4.015855312347412,
+ "step": 1976
+ },
+ {
+ "epoch": 27.46113537117904,
+ "grad_norm": 0.009037062525749207,
+ "learning_rate": 0.0013474082684628798,
+ "loss": 4.024164199829102,
+ "step": 1977
+ },
+ {
+ "epoch": 27.475109170305675,
+ "grad_norm": 0.009094092063605785,
+ "learning_rate": 0.0013460839189648028,
+ "loss": 4.037765979766846,
+ "step": 1978
+ },
+ {
+ "epoch": 27.489082969432314,
+ "grad_norm": 0.008230654522776604,
+ "learning_rate": 0.0013447597055593612,
+ "loss": 4.006317615509033,
+ "step": 1979
+ },
+ {
+ "epoch": 27.503056768558952,
+ "grad_norm": 0.008491289801895618,
+ "learning_rate": 0.0013434356293134501,
+ "loss": 4.042603969573975,
+ "step": 1980
+ },
+ {
+ "epoch": 27.51703056768559,
+ "grad_norm": 0.008907693438231945,
+ "learning_rate": 0.0013421116912938533,
+ "loss": 4.008098602294922,
+ "step": 1981
+ },
+ {
+ "epoch": 27.531004366812226,
+ "grad_norm": 0.008897818624973297,
+ "learning_rate": 0.0013407878925672438,
+ "loss": 3.9892945289611816,
+ "step": 1982
+ },
+ {
+ "epoch": 27.544978165938865,
+ "grad_norm": 0.00957555789500475,
+ "learning_rate": 0.0013394642342001832,
+ "loss": 4.046009063720703,
+ "step": 1983
+ },
+ {
+ "epoch": 27.558951965065503,
+ "grad_norm": 0.009948568418622017,
+ "learning_rate": 0.0013381407172591185,
+ "loss": 3.9760642051696777,
+ "step": 1984
+ },
+ {
+ "epoch": 27.57292576419214,
+ "grad_norm": 0.01014955248683691,
+ "learning_rate": 0.0013368173428103835,
+ "loss": 4.023215293884277,
+ "step": 1985
+ },
+ {
+ "epoch": 27.586899563318777,
+ "grad_norm": 0.008578895591199398,
+ "learning_rate": 0.0013354941119201966,
+ "loss": 4.002819061279297,
+ "step": 1986
+ },
+ {
+ "epoch": 27.600873362445416,
+ "grad_norm": 0.008729875087738037,
+ "learning_rate": 0.001334171025654662,
+ "loss": 4.0340657234191895,
+ "step": 1987
+ },
+ {
+ "epoch": 27.61484716157205,
+ "grad_norm": 0.009556280449032784,
+ "learning_rate": 0.0013328480850797655,
+ "loss": 3.983076333999634,
+ "step": 1988
+ },
+ {
+ "epoch": 27.62882096069869,
+ "grad_norm": 0.009733820334076881,
+ "learning_rate": 0.0013315252912613766,
+ "loss": 4.066578388214111,
+ "step": 1989
+ },
+ {
+ "epoch": 27.64279475982533,
+ "grad_norm": 0.009025252424180508,
+ "learning_rate": 0.0013302026452652474,
+ "loss": 3.9980697631835938,
+ "step": 1990
+ },
+ {
+ "epoch": 27.656768558951963,
+ "grad_norm": 0.008485320024192333,
+ "learning_rate": 0.001328880148157008,
+ "loss": 4.018826961517334,
+ "step": 1991
+ },
+ {
+ "epoch": 27.670742358078602,
+ "grad_norm": 0.009539696387946606,
+ "learning_rate": 0.0013275578010021726,
+ "loss": 4.011737823486328,
+ "step": 1992
+ },
+ {
+ "epoch": 27.68471615720524,
+ "grad_norm": 0.010881016030907631,
+ "learning_rate": 0.0013262356048661305,
+ "loss": 4.007291793823242,
+ "step": 1993
+ },
+ {
+ "epoch": 27.69868995633188,
+ "grad_norm": 0.010362236760556698,
+ "learning_rate": 0.0013249135608141523,
+ "loss": 3.9929394721984863,
+ "step": 1994
+ },
+ {
+ "epoch": 27.712663755458514,
+ "grad_norm": 0.010296608321368694,
+ "learning_rate": 0.0013235916699113849,
+ "loss": 4.0015130043029785,
+ "step": 1995
+ },
+ {
+ "epoch": 27.726637554585153,
+ "grad_norm": 0.011211174540221691,
+ "learning_rate": 0.0013222699332228517,
+ "loss": 4.005324363708496,
+ "step": 1996
+ },
+ {
+ "epoch": 27.74061135371179,
+ "grad_norm": 0.009743994101881981,
+ "learning_rate": 0.0013209483518134525,
+ "loss": 4.0278143882751465,
+ "step": 1997
+ },
+ {
+ "epoch": 27.754585152838427,
+ "grad_norm": 0.008563914336264133,
+ "learning_rate": 0.0013196269267479607,
+ "loss": 3.9997801780700684,
+ "step": 1998
+ },
+ {
+ "epoch": 27.768558951965066,
+ "grad_norm": 0.009078902192413807,
+ "learning_rate": 0.0013183056590910255,
+ "loss": 4.026200294494629,
+ "step": 1999
+ },
+ {
+ "epoch": 27.782532751091704,
+ "grad_norm": 0.009618895128369331,
+ "learning_rate": 0.0013169845499071675,
+ "loss": 4.027050018310547,
+ "step": 2000
+ },
+ {
+ "epoch": 27.79650655021834,
+ "grad_norm": 0.008920644409954548,
+ "learning_rate": 0.0013156636002607822,
+ "loss": 4.000490665435791,
+ "step": 2001
+ },
+ {
+ "epoch": 27.810480349344978,
+ "grad_norm": 0.008849202655255795,
+ "learning_rate": 0.0013143428112161323,
+ "loss": 4.044985294342041,
+ "step": 2002
+ },
+ {
+ "epoch": 27.824454148471617,
+ "grad_norm": 0.009303621016442776,
+ "learning_rate": 0.001313022183837356,
+ "loss": 4.046281814575195,
+ "step": 2003
+ },
+ {
+ "epoch": 27.83842794759825,
+ "grad_norm": 0.01076227892190218,
+ "learning_rate": 0.0013117017191884579,
+ "loss": 4.030576705932617,
+ "step": 2004
+ },
+ {
+ "epoch": 27.85240174672489,
+ "grad_norm": 0.010400841943919659,
+ "learning_rate": 0.0013103814183333127,
+ "loss": 3.980081558227539,
+ "step": 2005
+ },
+ {
+ "epoch": 27.86637554585153,
+ "grad_norm": 0.009459166787564754,
+ "learning_rate": 0.001309061282335663,
+ "loss": 4.014824867248535,
+ "step": 2006
+ },
+ {
+ "epoch": 27.880349344978168,
+ "grad_norm": 0.009054969064891338,
+ "learning_rate": 0.001307741312259118,
+ "loss": 4.0485944747924805,
+ "step": 2007
+ },
+ {
+ "epoch": 27.894323144104803,
+ "grad_norm": 0.009003991261124611,
+ "learning_rate": 0.0013064215091671545,
+ "loss": 3.9990320205688477,
+ "step": 2008
+ },
+ {
+ "epoch": 27.90829694323144,
+ "grad_norm": 0.008959448896348476,
+ "learning_rate": 0.001305101874123113,
+ "loss": 4.038620948791504,
+ "step": 2009
+ },
+ {
+ "epoch": 27.92227074235808,
+ "grad_norm": 0.009102243930101395,
+ "learning_rate": 0.0013037824081902015,
+ "loss": 4.030906677246094,
+ "step": 2010
+ },
+ {
+ "epoch": 27.936244541484715,
+ "grad_norm": 0.008599337190389633,
+ "learning_rate": 0.0013024631124314877,
+ "loss": 4.02648401260376,
+ "step": 2011
+ },
+ {
+ "epoch": 27.950218340611354,
+ "grad_norm": 0.008771977387368679,
+ "learning_rate": 0.0013011439879099056,
+ "loss": 4.027151107788086,
+ "step": 2012
+ },
+ {
+ "epoch": 27.964192139737992,
+ "grad_norm": 0.009082595817744732,
+ "learning_rate": 0.00129982503568825,
+ "loss": 4.0194478034973145,
+ "step": 2013
+ },
+ {
+ "epoch": 27.978165938864628,
+ "grad_norm": 0.009025354869663715,
+ "learning_rate": 0.0012985062568291763,
+ "loss": 4.044382095336914,
+ "step": 2014
+ },
+ {
+ "epoch": 27.992139737991266,
+ "grad_norm": 0.00875102635473013,
+ "learning_rate": 0.0012971876523952014,
+ "loss": 4.0257568359375,
+ "step": 2015
+ },
+ {
+ "epoch": 28.0,
+ "grad_norm": 0.009312704205513,
+ "learning_rate": 0.0012958692234487006,
+ "loss": 3.9964494705200195,
+ "step": 2016
+ },
+ {
+ "epoch": 28.0,
+ "eval_loss": 4.431707859039307,
+ "eval_runtime": 50.7311,
+ "eval_samples_per_second": 48.136,
+ "eval_steps_per_second": 1.518,
+ "step": 2016
+ },
+ {
+ "epoch": 28.01397379912664,
+ "grad_norm": 0.00992550328373909,
+ "learning_rate": 0.0012945509710519088,
+ "loss": 3.9253792762756348,
+ "step": 2017
+ },
+ {
+ "epoch": 28.027947598253274,
+ "grad_norm": 0.009920453652739525,
+ "learning_rate": 0.0012932328962669173,
+ "loss": 3.9438819885253906,
+ "step": 2018
+ },
+ {
+ "epoch": 28.041921397379912,
+ "grad_norm": 0.009437218308448792,
+ "learning_rate": 0.0012919150001556756,
+ "loss": 3.960963249206543,
+ "step": 2019
+ },
+ {
+ "epoch": 28.05589519650655,
+ "grad_norm": 0.009616893716156483,
+ "learning_rate": 0.0012905972837799896,
+ "loss": 3.969301223754883,
+ "step": 2020
+ },
+ {
+ "epoch": 28.069868995633186,
+ "grad_norm": 0.009107022546231747,
+ "learning_rate": 0.0012892797482015185,
+ "loss": 3.9626078605651855,
+ "step": 2021
+ },
+ {
+ "epoch": 28.083842794759825,
+ "grad_norm": 0.011317167431116104,
+ "learning_rate": 0.0012879623944817772,
+ "loss": 3.946483850479126,
+ "step": 2022
+ },
+ {
+ "epoch": 28.097816593886463,
+ "grad_norm": 0.010536770336329937,
+ "learning_rate": 0.001286645223682134,
+ "loss": 3.959566116333008,
+ "step": 2023
+ },
+ {
+ "epoch": 28.111790393013102,
+ "grad_norm": 0.009349033236503601,
+ "learning_rate": 0.0012853282368638097,
+ "loss": 3.964435338973999,
+ "step": 2024
+ },
+ {
+ "epoch": 28.125764192139737,
+ "grad_norm": 0.009950080886483192,
+ "learning_rate": 0.0012840114350878762,
+ "loss": 3.9729413986206055,
+ "step": 2025
+ },
+ {
+ "epoch": 28.139737991266376,
+ "grad_norm": 0.010833023115992546,
+ "learning_rate": 0.001282694819415257,
+ "loss": 4.006421089172363,
+ "step": 2026
+ },
+ {
+ "epoch": 28.153711790393015,
+ "grad_norm": 0.011538525111973286,
+ "learning_rate": 0.0012813783909067266,
+ "loss": 3.9558355808258057,
+ "step": 2027
+ },
+ {
+ "epoch": 28.16768558951965,
+ "grad_norm": 0.010671315714716911,
+ "learning_rate": 0.0012800621506229065,
+ "loss": 4.0052337646484375,
+ "step": 2028
+ },
+ {
+ "epoch": 28.18165938864629,
+ "grad_norm": 0.010835555382072926,
+ "learning_rate": 0.0012787460996242686,
+ "loss": 3.958641767501831,
+ "step": 2029
+ },
+ {
+ "epoch": 28.195633187772927,
+ "grad_norm": 0.01082367543131113,
+ "learning_rate": 0.0012774302389711304,
+ "loss": 3.9818902015686035,
+ "step": 2030
+ },
+ {
+ "epoch": 28.209606986899562,
+ "grad_norm": 0.009194346144795418,
+ "learning_rate": 0.0012761145697236583,
+ "loss": 3.9688854217529297,
+ "step": 2031
+ },
+ {
+ "epoch": 28.2235807860262,
+ "grad_norm": 0.007859935984015465,
+ "learning_rate": 0.001274799092941862,
+ "loss": 3.9730639457702637,
+ "step": 2032
+ },
+ {
+ "epoch": 28.23755458515284,
+ "grad_norm": 0.008534743450582027,
+ "learning_rate": 0.0012734838096855978,
+ "loss": 3.975003719329834,
+ "step": 2033
+ },
+ {
+ "epoch": 28.251528384279474,
+ "grad_norm": 0.009369445964694023,
+ "learning_rate": 0.0012721687210145666,
+ "loss": 3.972907304763794,
+ "step": 2034
+ },
+ {
+ "epoch": 28.265502183406113,
+ "grad_norm": 0.0090147340670228,
+ "learning_rate": 0.0012708538279883098,
+ "loss": 3.964256525039673,
+ "step": 2035
+ },
+ {
+ "epoch": 28.27947598253275,
+ "grad_norm": 0.007745139300823212,
+ "learning_rate": 0.0012695391316662151,
+ "loss": 3.9504833221435547,
+ "step": 2036
+ },
+ {
+ "epoch": 28.29344978165939,
+ "grad_norm": 0.008920101448893547,
+ "learning_rate": 0.0012682246331075075,
+ "loss": 3.9735097885131836,
+ "step": 2037
+ },
+ {
+ "epoch": 28.307423580786025,
+ "grad_norm": 0.008987226523458958,
+ "learning_rate": 0.0012669103333712568,
+ "loss": 3.9482927322387695,
+ "step": 2038
+ },
+ {
+ "epoch": 28.321397379912664,
+ "grad_norm": 0.00841289572417736,
+ "learning_rate": 0.0012655962335163685,
+ "loss": 3.965221881866455,
+ "step": 2039
+ },
+ {
+ "epoch": 28.335371179039303,
+ "grad_norm": 0.008016116917133331,
+ "learning_rate": 0.0012642823346015902,
+ "loss": 3.9390206336975098,
+ "step": 2040
+ },
+ {
+ "epoch": 28.349344978165938,
+ "grad_norm": 0.008470152504742146,
+ "learning_rate": 0.0012629686376855068,
+ "loss": 3.9257445335388184,
+ "step": 2041
+ },
+ {
+ "epoch": 28.363318777292577,
+ "grad_norm": 0.008637234568595886,
+ "learning_rate": 0.0012616551438265394,
+ "loss": 3.9999356269836426,
+ "step": 2042
+ },
+ {
+ "epoch": 28.377292576419215,
+ "grad_norm": 0.008036923594772816,
+ "learning_rate": 0.0012603418540829473,
+ "loss": 3.9755172729492188,
+ "step": 2043
+ },
+ {
+ "epoch": 28.39126637554585,
+ "grad_norm": 0.008908843621611595,
+ "learning_rate": 0.0012590287695128232,
+ "loss": 3.9248545169830322,
+ "step": 2044
+ },
+ {
+ "epoch": 28.40524017467249,
+ "grad_norm": 0.008743193000555038,
+ "learning_rate": 0.001257715891174096,
+ "loss": 3.9619390964508057,
+ "step": 2045
+ },
+ {
+ "epoch": 28.419213973799128,
+ "grad_norm": 0.00834199320524931,
+ "learning_rate": 0.001256403220124528,
+ "loss": 3.961569309234619,
+ "step": 2046
+ },
+ {
+ "epoch": 28.433187772925763,
+ "grad_norm": 0.009148893877863884,
+ "learning_rate": 0.001255090757421715,
+ "loss": 3.972104549407959,
+ "step": 2047
+ },
+ {
+ "epoch": 28.4471615720524,
+ "grad_norm": 0.00925084762275219,
+ "learning_rate": 0.0012537785041230843,
+ "loss": 4.000775337219238,
+ "step": 2048
+ },
+ {
+ "epoch": 28.46113537117904,
+ "grad_norm": 0.008659154176712036,
+ "learning_rate": 0.0012524664612858947,
+ "loss": 3.954564332962036,
+ "step": 2049
+ },
+ {
+ "epoch": 28.475109170305675,
+ "grad_norm": 0.007787859998643398,
+ "learning_rate": 0.0012511546299672355,
+ "loss": 3.9498860836029053,
+ "step": 2050
+ },
+ {
+ "epoch": 28.489082969432314,
+ "grad_norm": 0.00817918125540018,
+ "learning_rate": 0.0012498430112240246,
+ "loss": 4.003897666931152,
+ "step": 2051
+ },
+ {
+ "epoch": 28.503056768558952,
+ "grad_norm": 0.009282548911869526,
+ "learning_rate": 0.001248531606113011,
+ "loss": 3.9621925354003906,
+ "step": 2052
+ },
+ {
+ "epoch": 28.51703056768559,
+ "grad_norm": 0.009375610388815403,
+ "learning_rate": 0.001247220415690769,
+ "loss": 3.9977524280548096,
+ "step": 2053
+ },
+ {
+ "epoch": 28.531004366812226,
+ "grad_norm": 0.009650378488004208,
+ "learning_rate": 0.001245909441013701,
+ "loss": 3.9882829189300537,
+ "step": 2054
+ },
+ {
+ "epoch": 28.544978165938865,
+ "grad_norm": 0.009350919164717197,
+ "learning_rate": 0.0012445986831380366,
+ "loss": 4.019596099853516,
+ "step": 2055
+ },
+ {
+ "epoch": 28.558951965065503,
+ "grad_norm": 0.008726303465664387,
+ "learning_rate": 0.0012432881431198289,
+ "loss": 3.9933621883392334,
+ "step": 2056
+ },
+ {
+ "epoch": 28.57292576419214,
+ "grad_norm": 0.00829484686255455,
+ "learning_rate": 0.0012419778220149568,
+ "loss": 4.001237869262695,
+ "step": 2057
+ },
+ {
+ "epoch": 28.586899563318777,
+ "grad_norm": 0.008744452148675919,
+ "learning_rate": 0.0012406677208791218,
+ "loss": 3.9323887825012207,
+ "step": 2058
+ },
+ {
+ "epoch": 28.600873362445416,
+ "grad_norm": 0.008170492947101593,
+ "learning_rate": 0.0012393578407678496,
+ "loss": 3.925581455230713,
+ "step": 2059
+ },
+ {
+ "epoch": 28.61484716157205,
+ "grad_norm": 0.007843422703444958,
+ "learning_rate": 0.001238048182736486,
+ "loss": 4.001023769378662,
+ "step": 2060
+ },
+ {
+ "epoch": 28.62882096069869,
+ "grad_norm": 0.008914309553802013,
+ "learning_rate": 0.0012367387478401987,
+ "loss": 3.9804537296295166,
+ "step": 2061
+ },
+ {
+ "epoch": 28.64279475982533,
+ "grad_norm": 0.009736751206219196,
+ "learning_rate": 0.0012354295371339774,
+ "loss": 3.9831767082214355,
+ "step": 2062
+ },
+ {
+ "epoch": 28.656768558951963,
+ "grad_norm": 0.010061381384730339,
+ "learning_rate": 0.0012341205516726273,
+ "loss": 3.9462435245513916,
+ "step": 2063
+ },
+ {
+ "epoch": 28.670742358078602,
+ "grad_norm": 0.010350692085921764,
+ "learning_rate": 0.0012328117925107765,
+ "loss": 4.012863636016846,
+ "step": 2064
+ },
+ {
+ "epoch": 28.68471615720524,
+ "grad_norm": 0.01039902213960886,
+ "learning_rate": 0.0012315032607028667,
+ "loss": 3.9811301231384277,
+ "step": 2065
+ },
+ {
+ "epoch": 28.69868995633188,
+ "grad_norm": 0.011423371732234955,
+ "learning_rate": 0.0012301949573031595,
+ "loss": 3.9773542881011963,
+ "step": 2066
+ },
+ {
+ "epoch": 28.712663755458514,
+ "grad_norm": 0.012449614703655243,
+ "learning_rate": 0.001228886883365731,
+ "loss": 3.969600200653076,
+ "step": 2067
+ },
+ {
+ "epoch": 28.726637554585153,
+ "grad_norm": 0.010361429303884506,
+ "learning_rate": 0.0012275790399444726,
+ "loss": 4.035360336303711,
+ "step": 2068
+ },
+ {
+ "epoch": 28.74061135371179,
+ "grad_norm": 0.009603082202374935,
+ "learning_rate": 0.0012262714280930904,
+ "loss": 3.9840846061706543,
+ "step": 2069
+ },
+ {
+ "epoch": 28.754585152838427,
+ "grad_norm": 0.009183845482766628,
+ "learning_rate": 0.0012249640488651036,
+ "loss": 4.031542778015137,
+ "step": 2070
+ },
+ {
+ "epoch": 28.768558951965066,
+ "grad_norm": 0.009694446809589863,
+ "learning_rate": 0.0012236569033138441,
+ "loss": 3.9935057163238525,
+ "step": 2071
+ },
+ {
+ "epoch": 28.782532751091704,
+ "grad_norm": 0.0097816726192832,
+ "learning_rate": 0.0012223499924924545,
+ "loss": 3.9851176738739014,
+ "step": 2072
+ },
+ {
+ "epoch": 28.79650655021834,
+ "grad_norm": 0.009413184598088264,
+ "learning_rate": 0.001221043317453891,
+ "loss": 3.9642205238342285,
+ "step": 2073
+ },
+ {
+ "epoch": 28.810480349344978,
+ "grad_norm": 0.009960711002349854,
+ "learning_rate": 0.0012197368792509166,
+ "loss": 3.9373817443847656,
+ "step": 2074
+ },
+ {
+ "epoch": 28.824454148471617,
+ "grad_norm": 0.010124670341610909,
+ "learning_rate": 0.0012184306789361059,
+ "loss": 3.958115577697754,
+ "step": 2075
+ },
+ {
+ "epoch": 28.83842794759825,
+ "grad_norm": 0.011087656021118164,
+ "learning_rate": 0.00121712471756184,
+ "loss": 3.9607057571411133,
+ "step": 2076
+ },
+ {
+ "epoch": 28.85240174672489,
+ "grad_norm": 0.009812800213694572,
+ "learning_rate": 0.001215818996180309,
+ "loss": 4.00722599029541,
+ "step": 2077
+ },
+ {
+ "epoch": 28.86637554585153,
+ "grad_norm": 0.010305573232471943,
+ "learning_rate": 0.0012145135158435094,
+ "loss": 3.98728084564209,
+ "step": 2078
+ },
+ {
+ "epoch": 28.880349344978168,
+ "grad_norm": 0.01019787136465311,
+ "learning_rate": 0.0012132082776032422,
+ "loss": 3.981172800064087,
+ "step": 2079
+ },
+ {
+ "epoch": 28.894323144104803,
+ "grad_norm": 0.009260181337594986,
+ "learning_rate": 0.001211903282511115,
+ "loss": 4.034339904785156,
+ "step": 2080
+ },
+ {
+ "epoch": 28.90829694323144,
+ "grad_norm": 0.00972998607903719,
+ "learning_rate": 0.001210598531618538,
+ "loss": 4.022885322570801,
+ "step": 2081
+ },
+ {
+ "epoch": 28.92227074235808,
+ "grad_norm": 0.009853586554527283,
+ "learning_rate": 0.0012092940259767267,
+ "loss": 4.000235080718994,
+ "step": 2082
+ },
+ {
+ "epoch": 28.936244541484715,
+ "grad_norm": 0.00961449183523655,
+ "learning_rate": 0.0012079897666366962,
+ "loss": 3.9951353073120117,
+ "step": 2083
+ },
+ {
+ "epoch": 28.950218340611354,
+ "grad_norm": 0.010257176123559475,
+ "learning_rate": 0.0012066857546492657,
+ "loss": 3.9801371097564697,
+ "step": 2084
+ },
+ {
+ "epoch": 28.964192139737992,
+ "grad_norm": 0.009443561546504498,
+ "learning_rate": 0.001205381991065054,
+ "loss": 3.9596502780914307,
+ "step": 2085
+ },
+ {
+ "epoch": 28.978165938864628,
+ "grad_norm": 0.010389018803834915,
+ "learning_rate": 0.0012040784769344795,
+ "loss": 3.980527639389038,
+ "step": 2086
+ },
+ {
+ "epoch": 28.992139737991266,
+ "grad_norm": 0.010322567075490952,
+ "learning_rate": 0.0012027752133077607,
+ "loss": 3.996258497238159,
+ "step": 2087
+ },
+ {
+ "epoch": 29.0,
+ "grad_norm": 0.009987728670239449,
+ "learning_rate": 0.0012014722012349125,
+ "loss": 3.9493203163146973,
+ "step": 2088
+ },
+ {
+ "epoch": 29.0,
+ "eval_loss": 4.421086311340332,
+ "eval_runtime": 52.5623,
+ "eval_samples_per_second": 46.459,
+ "eval_steps_per_second": 1.465,
+ "step": 2088
+ },
+ {
+ "epoch": 29.01397379912664,
+ "grad_norm": 0.009405115619301796,
+ "learning_rate": 0.0012001694417657493,
+ "loss": 3.927506923675537,
+ "step": 2089
+ },
+ {
+ "epoch": 29.027947598253274,
+ "grad_norm": 0.009974643588066101,
+ "learning_rate": 0.00119886693594988,
+ "loss": 3.9373550415039062,
+ "step": 2090
+ },
+ {
+ "epoch": 29.041921397379912,
+ "grad_norm": 0.011331096291542053,
+ "learning_rate": 0.0011975646848367098,
+ "loss": 3.935594081878662,
+ "step": 2091
+ },
+ {
+ "epoch": 29.05589519650655,
+ "grad_norm": 0.010538448579609394,
+ "learning_rate": 0.0011962626894754405,
+ "loss": 3.911579132080078,
+ "step": 2092
+ },
+ {
+ "epoch": 29.069868995633186,
+ "grad_norm": 0.00989572424441576,
+ "learning_rate": 0.0011949609509150645,
+ "loss": 3.92525577545166,
+ "step": 2093
+ },
+ {
+ "epoch": 29.083842794759825,
+ "grad_norm": 0.010533696971833706,
+ "learning_rate": 0.0011936594702043704,
+ "loss": 3.898075580596924,
+ "step": 2094
+ },
+ {
+ "epoch": 29.097816593886463,
+ "grad_norm": 0.01051455456763506,
+ "learning_rate": 0.0011923582483919365,
+ "loss": 3.93463134765625,
+ "step": 2095
+ },
+ {
+ "epoch": 29.111790393013102,
+ "grad_norm": 0.009662075899541378,
+ "learning_rate": 0.0011910572865261348,
+ "loss": 3.898221731185913,
+ "step": 2096
+ },
+ {
+ "epoch": 29.125764192139737,
+ "grad_norm": 0.008874096907675266,
+ "learning_rate": 0.0011897565856551257,
+ "loss": 3.926787853240967,
+ "step": 2097
+ },
+ {
+ "epoch": 29.139737991266376,
+ "grad_norm": 0.008126807399094105,
+ "learning_rate": 0.0011884561468268606,
+ "loss": 3.9158413410186768,
+ "step": 2098
+ },
+ {
+ "epoch": 29.153711790393015,
+ "grad_norm": 0.007915553636848927,
+ "learning_rate": 0.001187155971089081,
+ "loss": 3.8891496658325195,
+ "step": 2099
+ },
+ {
+ "epoch": 29.16768558951965,
+ "grad_norm": 0.008312922902405262,
+ "learning_rate": 0.0011858560594893129,
+ "loss": 3.932219982147217,
+ "step": 2100
+ },
+ {
+ "epoch": 29.18165938864629,
+ "grad_norm": 0.008892328478395939,
+ "learning_rate": 0.0011845564130748734,
+ "loss": 3.8866336345672607,
+ "step": 2101
+ },
+ {
+ "epoch": 29.195633187772927,
+ "grad_norm": 0.009187526069581509,
+ "learning_rate": 0.0011832570328928628,
+ "loss": 3.9464707374572754,
+ "step": 2102
+ },
+ {
+ "epoch": 29.209606986899562,
+ "grad_norm": 0.008615494705736637,
+ "learning_rate": 0.0011819579199901693,
+ "loss": 3.9396629333496094,
+ "step": 2103
+ },
+ {
+ "epoch": 29.2235807860262,
+ "grad_norm": 0.008649308234453201,
+ "learning_rate": 0.0011806590754134636,
+ "loss": 3.9325060844421387,
+ "step": 2104
+ },
+ {
+ "epoch": 29.23755458515284,
+ "grad_norm": 0.008581082336604595,
+ "learning_rate": 0.0011793605002092021,
+ "loss": 3.9200944900512695,
+ "step": 2105
+ },
+ {
+ "epoch": 29.251528384279474,
+ "grad_norm": 0.008502699434757233,
+ "learning_rate": 0.0011780621954236238,
+ "loss": 3.9656100273132324,
+ "step": 2106
+ },
+ {
+ "epoch": 29.265502183406113,
+ "grad_norm": 0.0083576999604702,
+ "learning_rate": 0.0011767641621027477,
+ "loss": 3.9130215644836426,
+ "step": 2107
+ },
+ {
+ "epoch": 29.27947598253275,
+ "grad_norm": 0.008196269161999226,
+ "learning_rate": 0.0011754664012923785,
+ "loss": 3.942369222640991,
+ "step": 2108
+ },
+ {
+ "epoch": 29.29344978165939,
+ "grad_norm": 0.008857307955622673,
+ "learning_rate": 0.0011741689140380956,
+ "loss": 3.909585952758789,
+ "step": 2109
+ },
+ {
+ "epoch": 29.307423580786025,
+ "grad_norm": 0.008935349993407726,
+ "learning_rate": 0.0011728717013852635,
+ "loss": 3.9248366355895996,
+ "step": 2110
+ },
+ {
+ "epoch": 29.321397379912664,
+ "grad_norm": 0.008821200579404831,
+ "learning_rate": 0.0011715747643790211,
+ "loss": 3.9589288234710693,
+ "step": 2111
+ },
+ {
+ "epoch": 29.335371179039303,
+ "grad_norm": 0.00861942209303379,
+ "learning_rate": 0.0011702781040642885,
+ "loss": 3.9381728172302246,
+ "step": 2112
+ },
+ {
+ "epoch": 29.349344978165938,
+ "grad_norm": 0.009127574972808361,
+ "learning_rate": 0.0011689817214857613,
+ "loss": 3.897829532623291,
+ "step": 2113
+ },
+ {
+ "epoch": 29.363318777292577,
+ "grad_norm": 0.008008269593119621,
+ "learning_rate": 0.0011676856176879107,
+ "loss": 3.930056095123291,
+ "step": 2114
+ },
+ {
+ "epoch": 29.377292576419215,
+ "grad_norm": 0.010070208460092545,
+ "learning_rate": 0.0011663897937149853,
+ "loss": 3.892270803451538,
+ "step": 2115
+ },
+ {
+ "epoch": 29.39126637554585,
+ "grad_norm": 0.00903113093227148,
+ "learning_rate": 0.001165094250611006,
+ "loss": 3.9766440391540527,
+ "step": 2116
+ },
+ {
+ "epoch": 29.40524017467249,
+ "grad_norm": 0.008872415870428085,
+ "learning_rate": 0.0011637989894197692,
+ "loss": 3.9559712409973145,
+ "step": 2117
+ },
+ {
+ "epoch": 29.419213973799128,
+ "grad_norm": 0.009524816647171974,
+ "learning_rate": 0.0011625040111848427,
+ "loss": 3.9607882499694824,
+ "step": 2118
+ },
+ {
+ "epoch": 29.433187772925763,
+ "grad_norm": 0.009836779907345772,
+ "learning_rate": 0.0011612093169495678,
+ "loss": 3.9629592895507812,
+ "step": 2119
+ },
+ {
+ "epoch": 29.4471615720524,
+ "grad_norm": 0.009586540050804615,
+ "learning_rate": 0.0011599149077570561,
+ "loss": 3.95208740234375,
+ "step": 2120
+ },
+ {
+ "epoch": 29.46113537117904,
+ "grad_norm": 0.010495901107788086,
+ "learning_rate": 0.0011586207846501896,
+ "loss": 3.9439730644226074,
+ "step": 2121
+ },
+ {
+ "epoch": 29.475109170305675,
+ "grad_norm": 0.010346266441047192,
+ "learning_rate": 0.0011573269486716202,
+ "loss": 3.959299087524414,
+ "step": 2122
+ },
+ {
+ "epoch": 29.489082969432314,
+ "grad_norm": 0.010767634026706219,
+ "learning_rate": 0.0011560334008637674,
+ "loss": 3.900784969329834,
+ "step": 2123
+ },
+ {
+ "epoch": 29.503056768558952,
+ "grad_norm": 0.012122200801968575,
+ "learning_rate": 0.0011547401422688202,
+ "loss": 3.9588193893432617,
+ "step": 2124
+ },
+ {
+ "epoch": 29.51703056768559,
+ "grad_norm": 0.012914645485579967,
+ "learning_rate": 0.0011534471739287331,
+ "loss": 3.9278407096862793,
+ "step": 2125
+ },
+ {
+ "epoch": 29.531004366812226,
+ "grad_norm": 0.011785333044826984,
+ "learning_rate": 0.0011521544968852273,
+ "loss": 3.992793560028076,
+ "step": 2126
+ },
+ {
+ "epoch": 29.544978165938865,
+ "grad_norm": 0.008455781266093254,
+ "learning_rate": 0.0011508621121797899,
+ "loss": 3.96525239944458,
+ "step": 2127
+ },
+ {
+ "epoch": 29.558951965065503,
+ "grad_norm": 0.009655282832682133,
+ "learning_rate": 0.0011495700208536714,
+ "loss": 3.9346890449523926,
+ "step": 2128
+ },
+ {
+ "epoch": 29.57292576419214,
+ "grad_norm": 0.010537004098296165,
+ "learning_rate": 0.0011482782239478871,
+ "loss": 3.9534668922424316,
+ "step": 2129
+ },
+ {
+ "epoch": 29.586899563318777,
+ "grad_norm": 0.009438651613891125,
+ "learning_rate": 0.0011469867225032135,
+ "loss": 3.983487606048584,
+ "step": 2130
+ },
+ {
+ "epoch": 29.600873362445416,
+ "grad_norm": 0.010586646385490894,
+ "learning_rate": 0.0011456955175601908,
+ "loss": 3.939680576324463,
+ "step": 2131
+ },
+ {
+ "epoch": 29.61484716157205,
+ "grad_norm": 0.009082718752324581,
+ "learning_rate": 0.001144404610159119,
+ "loss": 3.9558515548706055,
+ "step": 2132
+ },
+ {
+ "epoch": 29.62882096069869,
+ "grad_norm": 0.009833602234721184,
+ "learning_rate": 0.0011431140013400591,
+ "loss": 3.9539828300476074,
+ "step": 2133
+ },
+ {
+ "epoch": 29.64279475982533,
+ "grad_norm": 0.009142170660197735,
+ "learning_rate": 0.0011418236921428313,
+ "loss": 3.9442849159240723,
+ "step": 2134
+ },
+ {
+ "epoch": 29.656768558951963,
+ "grad_norm": 0.008603853173553944,
+ "learning_rate": 0.0011405336836070137,
+ "loss": 3.9526913166046143,
+ "step": 2135
+ },
+ {
+ "epoch": 29.670742358078602,
+ "grad_norm": 0.008838302455842495,
+ "learning_rate": 0.0011392439767719446,
+ "loss": 3.973104953765869,
+ "step": 2136
+ },
+ {
+ "epoch": 29.68471615720524,
+ "grad_norm": 0.008867006748914719,
+ "learning_rate": 0.0011379545726767156,
+ "loss": 3.9832305908203125,
+ "step": 2137
+ },
+ {
+ "epoch": 29.69868995633188,
+ "grad_norm": 0.00971697736531496,
+ "learning_rate": 0.001136665472360178,
+ "loss": 3.9563937187194824,
+ "step": 2138
+ },
+ {
+ "epoch": 29.712663755458514,
+ "grad_norm": 0.00957175437361002,
+ "learning_rate": 0.001135376676860935,
+ "loss": 3.975194215774536,
+ "step": 2139
+ },
+ {
+ "epoch": 29.726637554585153,
+ "grad_norm": 0.008548660203814507,
+ "learning_rate": 0.001134088187217347,
+ "loss": 3.9839985370635986,
+ "step": 2140
+ },
+ {
+ "epoch": 29.74061135371179,
+ "grad_norm": 0.008556334301829338,
+ "learning_rate": 0.0011328000044675264,
+ "loss": 3.9225516319274902,
+ "step": 2141
+ },
+ {
+ "epoch": 29.754585152838427,
+ "grad_norm": 0.008440867997705936,
+ "learning_rate": 0.0011315121296493385,
+ "loss": 3.945028066635132,
+ "step": 2142
+ },
+ {
+ "epoch": 29.768558951965066,
+ "grad_norm": 0.009643333032727242,
+ "learning_rate": 0.0011302245638004017,
+ "loss": 3.943120002746582,
+ "step": 2143
+ },
+ {
+ "epoch": 29.782532751091704,
+ "grad_norm": 0.008832835592329502,
+ "learning_rate": 0.001128937307958083,
+ "loss": 3.9895920753479004,
+ "step": 2144
+ },
+ {
+ "epoch": 29.79650655021834,
+ "grad_norm": 0.008720474317669868,
+ "learning_rate": 0.0011276503631595036,
+ "loss": 3.941950798034668,
+ "step": 2145
+ },
+ {
+ "epoch": 29.810480349344978,
+ "grad_norm": 0.008099926635622978,
+ "learning_rate": 0.001126363730441529,
+ "loss": 3.9626870155334473,
+ "step": 2146
+ },
+ {
+ "epoch": 29.824454148471617,
+ "grad_norm": 0.008525107055902481,
+ "learning_rate": 0.0011250774108407778,
+ "loss": 3.950535774230957,
+ "step": 2147
+ },
+ {
+ "epoch": 29.83842794759825,
+ "grad_norm": 0.008998272009193897,
+ "learning_rate": 0.0011237914053936135,
+ "loss": 3.9460678100585938,
+ "step": 2148
+ },
+ {
+ "epoch": 29.85240174672489,
+ "grad_norm": 0.009387437254190445,
+ "learning_rate": 0.001122505715136148,
+ "loss": 3.926604986190796,
+ "step": 2149
+ },
+ {
+ "epoch": 29.86637554585153,
+ "grad_norm": 0.007804017513990402,
+ "learning_rate": 0.0011212203411042386,
+ "loss": 3.9894230365753174,
+ "step": 2150
+ },
+ {
+ "epoch": 29.880349344978168,
+ "grad_norm": 0.008564616553485394,
+ "learning_rate": 0.0011199352843334876,
+ "loss": 3.9339959621429443,
+ "step": 2151
+ },
+ {
+ "epoch": 29.894323144104803,
+ "grad_norm": 0.008726440370082855,
+ "learning_rate": 0.0011186505458592427,
+ "loss": 3.9867987632751465,
+ "step": 2152
+ },
+ {
+ "epoch": 29.90829694323144,
+ "grad_norm": 0.008074100129306316,
+ "learning_rate": 0.0011173661267165932,
+ "loss": 3.9492292404174805,
+ "step": 2153
+ },
+ {
+ "epoch": 29.92227074235808,
+ "grad_norm": 0.007688041776418686,
+ "learning_rate": 0.0011160820279403744,
+ "loss": 3.9425454139709473,
+ "step": 2154
+ },
+ {
+ "epoch": 29.936244541484715,
+ "grad_norm": 0.007808993570506573,
+ "learning_rate": 0.0011147982505651595,
+ "loss": 3.935114860534668,
+ "step": 2155
+ },
+ {
+ "epoch": 29.950218340611354,
+ "grad_norm": 0.009786820039153099,
+ "learning_rate": 0.0011135147956252659,
+ "loss": 3.959758758544922,
+ "step": 2156
+ },
+ {
+ "epoch": 29.964192139737992,
+ "grad_norm": 0.009487155824899673,
+ "learning_rate": 0.00111223166415475,
+ "loss": 3.989048480987549,
+ "step": 2157
+ },
+ {
+ "epoch": 29.978165938864628,
+ "grad_norm": 0.008407804183661938,
+ "learning_rate": 0.0011109488571874073,
+ "loss": 3.932347297668457,
+ "step": 2158
+ },
+ {
+ "epoch": 29.992139737991266,
+ "grad_norm": 0.007158380467444658,
+ "learning_rate": 0.0011096663757567723,
+ "loss": 3.9545154571533203,
+ "step": 2159
+ },
+ {
+ "epoch": 30.0,
+ "grad_norm": 0.008840581402182579,
+ "learning_rate": 0.001108384220896117,
+ "loss": 3.965984344482422,
+ "step": 2160
+ },
+ {
+ "epoch": 30.0,
+ "eval_loss": 4.409194469451904,
+ "eval_runtime": 51.5561,
+ "eval_samples_per_second": 47.366,
+ "eval_steps_per_second": 1.494,
+ "step": 2160
+ },
+ {
+ "epoch": 30.01397379912664,
+ "grad_norm": 0.010702896863222122,
+ "learning_rate": 0.0011071023936384508,
+ "loss": 3.8964462280273438,
+ "step": 2161
+ },
+ {
+ "epoch": 30.027947598253274,
+ "grad_norm": 0.01290619745850563,
+ "learning_rate": 0.0011058208950165183,
+ "loss": 3.8775625228881836,
+ "step": 2162
+ },
+ {
+ "epoch": 30.041921397379912,
+ "grad_norm": 0.01443667896091938,
+ "learning_rate": 0.0011045397260628,
+ "loss": 3.8755288124084473,
+ "step": 2163
+ },
+ {
+ "epoch": 30.05589519650655,
+ "grad_norm": 0.014411968179047108,
+ "learning_rate": 0.0011032588878095108,
+ "loss": 3.889918327331543,
+ "step": 2164
+ },
+ {
+ "epoch": 30.069868995633186,
+ "grad_norm": 0.013656266033649445,
+ "learning_rate": 0.0011019783812885986,
+ "loss": 3.926997661590576,
+ "step": 2165
+ },
+ {
+ "epoch": 30.083842794759825,
+ "grad_norm": 0.010661937296390533,
+ "learning_rate": 0.0011006982075317446,
+ "loss": 3.9226956367492676,
+ "step": 2166
+ },
+ {
+ "epoch": 30.097816593886463,
+ "grad_norm": 0.008968768641352654,
+ "learning_rate": 0.0010994183675703615,
+ "loss": 3.8932647705078125,
+ "step": 2167
+ },
+ {
+ "epoch": 30.111790393013102,
+ "grad_norm": 0.009992181323468685,
+ "learning_rate": 0.0010981388624355937,
+ "loss": 3.8920068740844727,
+ "step": 2168
+ },
+ {
+ "epoch": 30.125764192139737,
+ "grad_norm": 0.011034460738301277,
+ "learning_rate": 0.0010968596931583148,
+ "loss": 3.909073829650879,
+ "step": 2169
+ },
+ {
+ "epoch": 30.139737991266376,
+ "grad_norm": 0.011815821751952171,
+ "learning_rate": 0.0010955808607691284,
+ "loss": 3.9190006256103516,
+ "step": 2170
+ },
+ {
+ "epoch": 30.153711790393015,
+ "grad_norm": 0.012317490763962269,
+ "learning_rate": 0.0010943023662983674,
+ "loss": 3.8790552616119385,
+ "step": 2171
+ },
+ {
+ "epoch": 30.16768558951965,
+ "grad_norm": 0.010983958840370178,
+ "learning_rate": 0.0010930242107760903,
+ "loss": 3.8730852603912354,
+ "step": 2172
+ },
+ {
+ "epoch": 30.18165938864629,
+ "grad_norm": 0.008948802947998047,
+ "learning_rate": 0.0010917463952320856,
+ "loss": 3.923659563064575,
+ "step": 2173
+ },
+ {
+ "epoch": 30.195633187772927,
+ "grad_norm": 0.009568693116307259,
+ "learning_rate": 0.0010904689206958649,
+ "loss": 3.9100794792175293,
+ "step": 2174
+ },
+ {
+ "epoch": 30.209606986899562,
+ "grad_norm": 0.011862963438034058,
+ "learning_rate": 0.0010891917881966672,
+ "loss": 3.908093214035034,
+ "step": 2175
+ },
+ {
+ "epoch": 30.2235807860262,
+ "grad_norm": 0.010180995799601078,
+ "learning_rate": 0.0010879149987634545,
+ "loss": 3.900660514831543,
+ "step": 2176
+ },
+ {
+ "epoch": 30.23755458515284,
+ "grad_norm": 0.009372464381158352,
+ "learning_rate": 0.0010866385534249131,
+ "loss": 3.8950703144073486,
+ "step": 2177
+ },
+ {
+ "epoch": 30.251528384279474,
+ "grad_norm": 0.010695629753172398,
+ "learning_rate": 0.0010853624532094525,
+ "loss": 3.913530111312866,
+ "step": 2178
+ },
+ {
+ "epoch": 30.265502183406113,
+ "grad_norm": 0.011817459017038345,
+ "learning_rate": 0.0010840866991452024,
+ "loss": 3.90268611907959,
+ "step": 2179
+ },
+ {
+ "epoch": 30.27947598253275,
+ "grad_norm": 0.010686337016522884,
+ "learning_rate": 0.0010828112922600167,
+ "loss": 3.9141364097595215,
+ "step": 2180
+ },
+ {
+ "epoch": 30.29344978165939,
+ "grad_norm": 0.009550466202199459,
+ "learning_rate": 0.0010815362335814655,
+ "loss": 3.968468189239502,
+ "step": 2181
+ },
+ {
+ "epoch": 30.307423580786025,
+ "grad_norm": 0.009367208927869797,
+ "learning_rate": 0.0010802615241368423,
+ "loss": 3.8933968544006348,
+ "step": 2182
+ },
+ {
+ "epoch": 30.321397379912664,
+ "grad_norm": 0.009899397380650043,
+ "learning_rate": 0.0010789871649531562,
+ "loss": 3.936708450317383,
+ "step": 2183
+ },
+ {
+ "epoch": 30.335371179039303,
+ "grad_norm": 0.009181005880236626,
+ "learning_rate": 0.0010777131570571363,
+ "loss": 3.9370503425598145,
+ "step": 2184
+ },
+ {
+ "epoch": 30.349344978165938,
+ "grad_norm": 0.00799303874373436,
+ "learning_rate": 0.0010764395014752274,
+ "loss": 3.911050319671631,
+ "step": 2185
+ },
+ {
+ "epoch": 30.363318777292577,
+ "grad_norm": 0.008469315245747566,
+ "learning_rate": 0.0010751661992335908,
+ "loss": 3.9103240966796875,
+ "step": 2186
+ },
+ {
+ "epoch": 30.377292576419215,
+ "grad_norm": 0.009814459830522537,
+ "learning_rate": 0.0010738932513581034,
+ "loss": 3.881843090057373,
+ "step": 2187
+ },
+ {
+ "epoch": 30.39126637554585,
+ "grad_norm": 0.010244108736515045,
+ "learning_rate": 0.0010726206588743557,
+ "loss": 3.883141040802002,
+ "step": 2188
+ },
+ {
+ "epoch": 30.40524017467249,
+ "grad_norm": 0.008898732252418995,
+ "learning_rate": 0.001071348422807653,
+ "loss": 3.952347755432129,
+ "step": 2189
+ },
+ {
+ "epoch": 30.419213973799128,
+ "grad_norm": 0.00823870487511158,
+ "learning_rate": 0.0010700765441830121,
+ "loss": 3.932507038116455,
+ "step": 2190
+ },
+ {
+ "epoch": 30.433187772925763,
+ "grad_norm": 0.009857657365500927,
+ "learning_rate": 0.001068805024025164,
+ "loss": 3.932932138442993,
+ "step": 2191
+ },
+ {
+ "epoch": 30.4471615720524,
+ "grad_norm": 0.00929308496415615,
+ "learning_rate": 0.0010675338633585488,
+ "loss": 3.8971338272094727,
+ "step": 2192
+ },
+ {
+ "epoch": 30.46113537117904,
+ "grad_norm": 0.008667961694300175,
+ "learning_rate": 0.0010662630632073172,
+ "loss": 3.916931629180908,
+ "step": 2193
+ },
+ {
+ "epoch": 30.475109170305675,
+ "grad_norm": 0.009758287109434605,
+ "learning_rate": 0.001064992624595331,
+ "loss": 3.964726209640503,
+ "step": 2194
+ },
+ {
+ "epoch": 30.489082969432314,
+ "grad_norm": 0.009429254569113255,
+ "learning_rate": 0.0010637225485461586,
+ "loss": 3.917572498321533,
+ "step": 2195
+ },
+ {
+ "epoch": 30.503056768558952,
+ "grad_norm": 0.008303353562951088,
+ "learning_rate": 0.0010624528360830782,
+ "loss": 3.875241756439209,
+ "step": 2196
+ },
+ {
+ "epoch": 30.51703056768559,
+ "grad_norm": 0.008729013614356518,
+ "learning_rate": 0.001061183488229073,
+ "loss": 3.9691600799560547,
+ "step": 2197
+ },
+ {
+ "epoch": 30.531004366812226,
+ "grad_norm": 0.010239302180707455,
+ "learning_rate": 0.0010599145060068353,
+ "loss": 3.9053149223327637,
+ "step": 2198
+ },
+ {
+ "epoch": 30.544978165938865,
+ "grad_norm": 0.00932721421122551,
+ "learning_rate": 0.0010586458904387596,
+ "loss": 3.8593955039978027,
+ "step": 2199
+ },
+ {
+ "epoch": 30.558951965065503,
+ "grad_norm": 0.008196988143026829,
+ "learning_rate": 0.0010573776425469473,
+ "loss": 3.917983055114746,
+ "step": 2200
+ },
+ {
+ "epoch": 30.57292576419214,
+ "grad_norm": 0.008049141615629196,
+ "learning_rate": 0.001056109763353203,
+ "loss": 3.9337785243988037,
+ "step": 2201
+ },
+ {
+ "epoch": 30.586899563318777,
+ "grad_norm": 0.009256313554942608,
+ "learning_rate": 0.001054842253879034,
+ "loss": 3.937887668609619,
+ "step": 2202
+ },
+ {
+ "epoch": 30.600873362445416,
+ "grad_norm": 0.009292179718613625,
+ "learning_rate": 0.0010535751151456495,
+ "loss": 3.9203450679779053,
+ "step": 2203
+ },
+ {
+ "epoch": 30.61484716157205,
+ "grad_norm": 0.009142161346971989,
+ "learning_rate": 0.0010523083481739602,
+ "loss": 3.9027457237243652,
+ "step": 2204
+ },
+ {
+ "epoch": 30.62882096069869,
+ "grad_norm": 0.009498288854956627,
+ "learning_rate": 0.0010510419539845777,
+ "loss": 3.917316436767578,
+ "step": 2205
+ },
+ {
+ "epoch": 30.64279475982533,
+ "grad_norm": 0.00849096104502678,
+ "learning_rate": 0.001049775933597813,
+ "loss": 3.9229235649108887,
+ "step": 2206
+ },
+ {
+ "epoch": 30.656768558951963,
+ "grad_norm": 0.008343606255948544,
+ "learning_rate": 0.0010485102880336747,
+ "loss": 3.9138073921203613,
+ "step": 2207
+ },
+ {
+ "epoch": 30.670742358078602,
+ "grad_norm": 0.010342134162783623,
+ "learning_rate": 0.0010472450183118727,
+ "loss": 3.8932929039001465,
+ "step": 2208
+ },
+ {
+ "epoch": 30.68471615720524,
+ "grad_norm": 0.010298306122422218,
+ "learning_rate": 0.00104598012545181,
+ "loss": 3.9112677574157715,
+ "step": 2209
+ },
+ {
+ "epoch": 30.69868995633188,
+ "grad_norm": 0.00827560480684042,
+ "learning_rate": 0.001044715610472589,
+ "loss": 3.894556760787964,
+ "step": 2210
+ },
+ {
+ "epoch": 30.712663755458514,
+ "grad_norm": 0.008993102237582207,
+ "learning_rate": 0.0010434514743930064,
+ "loss": 3.905391216278076,
+ "step": 2211
+ },
+ {
+ "epoch": 30.726637554585153,
+ "grad_norm": 0.008632538840174675,
+ "learning_rate": 0.0010421877182315536,
+ "loss": 3.896714210510254,
+ "step": 2212
+ },
+ {
+ "epoch": 30.74061135371179,
+ "grad_norm": 0.008542424067854881,
+ "learning_rate": 0.0010409243430064162,
+ "loss": 3.9145569801330566,
+ "step": 2213
+ },
+ {
+ "epoch": 30.754585152838427,
+ "grad_norm": 0.008970450609922409,
+ "learning_rate": 0.0010396613497354725,
+ "loss": 3.9170360565185547,
+ "step": 2214
+ },
+ {
+ "epoch": 30.768558951965066,
+ "grad_norm": 0.008744142949581146,
+ "learning_rate": 0.0010383987394362942,
+ "loss": 3.927412509918213,
+ "step": 2215
+ },
+ {
+ "epoch": 30.782532751091704,
+ "grad_norm": 0.008229847066104412,
+ "learning_rate": 0.0010371365131261424,
+ "loss": 3.928352117538452,
+ "step": 2216
+ },
+ {
+ "epoch": 30.79650655021834,
+ "grad_norm": 0.00818693358451128,
+ "learning_rate": 0.0010358746718219717,
+ "loss": 3.9219343662261963,
+ "step": 2217
+ },
+ {
+ "epoch": 30.810480349344978,
+ "grad_norm": 0.008848588913679123,
+ "learning_rate": 0.001034613216540423,
+ "loss": 3.9346699714660645,
+ "step": 2218
+ },
+ {
+ "epoch": 30.824454148471617,
+ "grad_norm": 0.00834780652076006,
+ "learning_rate": 0.0010333521482978295,
+ "loss": 3.9144492149353027,
+ "step": 2219
+ },
+ {
+ "epoch": 30.83842794759825,
+ "grad_norm": 0.00940527394413948,
+ "learning_rate": 0.0010320914681102098,
+ "loss": 3.951573371887207,
+ "step": 2220
+ },
+ {
+ "epoch": 30.85240174672489,
+ "grad_norm": 0.008884521201252937,
+ "learning_rate": 0.0010308311769932725,
+ "loss": 3.899625778198242,
+ "step": 2221
+ },
+ {
+ "epoch": 30.86637554585153,
+ "grad_norm": 0.007863261736929417,
+ "learning_rate": 0.0010295712759624107,
+ "loss": 3.9321727752685547,
+ "step": 2222
+ },
+ {
+ "epoch": 30.880349344978168,
+ "grad_norm": 0.008651153184473515,
+ "learning_rate": 0.0010283117660327035,
+ "loss": 3.9030094146728516,
+ "step": 2223
+ },
+ {
+ "epoch": 30.894323144104803,
+ "grad_norm": 0.00822074618190527,
+ "learning_rate": 0.0010270526482189163,
+ "loss": 3.939466953277588,
+ "step": 2224
+ },
+ {
+ "epoch": 30.90829694323144,
+ "grad_norm": 0.008701153099536896,
+ "learning_rate": 0.0010257939235354964,
+ "loss": 3.9099626541137695,
+ "step": 2225
+ },
+ {
+ "epoch": 30.92227074235808,
+ "grad_norm": 0.008418620564043522,
+ "learning_rate": 0.0010245355929965774,
+ "loss": 3.914116382598877,
+ "step": 2226
+ },
+ {
+ "epoch": 30.936244541484715,
+ "grad_norm": 0.008339700289070606,
+ "learning_rate": 0.001023277657615971,
+ "loss": 3.9111878871917725,
+ "step": 2227
+ },
+ {
+ "epoch": 30.950218340611354,
+ "grad_norm": 0.008304647170007229,
+ "learning_rate": 0.0010220201184071747,
+ "loss": 3.950974464416504,
+ "step": 2228
+ },
+ {
+ "epoch": 30.964192139737992,
+ "grad_norm": 0.008436229079961777,
+ "learning_rate": 0.001020762976383365,
+ "loss": 3.9631457328796387,
+ "step": 2229
+ },
+ {
+ "epoch": 30.978165938864628,
+ "grad_norm": 0.0074487063102424145,
+ "learning_rate": 0.001019506232557398,
+ "loss": 3.902984619140625,
+ "step": 2230
+ },
+ {
+ "epoch": 30.992139737991266,
+ "grad_norm": 0.008736515417695045,
+ "learning_rate": 0.0010182498879418096,
+ "loss": 3.9268949031829834,
+ "step": 2231
+ },
+ {
+ "epoch": 31.0,
+ "grad_norm": 0.009666468016803265,
+ "learning_rate": 0.0010169939435488145,
+ "loss": 3.9274301528930664,
+ "step": 2232
+ },
+ {
+ "epoch": 31.0,
+ "eval_loss": 4.395970821380615,
+ "eval_runtime": 53.7848,
+ "eval_samples_per_second": 45.403,
+ "eval_steps_per_second": 1.432,
+ "step": 2232
+ },
+ {
+ "epoch": 31.01397379912664,
+ "grad_norm": 0.009094545617699623,
+ "learning_rate": 0.001015738400390304,
+ "loss": 3.866425037384033,
+ "step": 2233
+ },
+ {
+ "epoch": 31.027947598253274,
+ "grad_norm": 0.009273733012378216,
+ "learning_rate": 0.0010144832594778462,
+ "loss": 3.8569841384887695,
+ "step": 2234
+ },
+ {
+ "epoch": 31.041921397379912,
+ "grad_norm": 0.009205271489918232,
+ "learning_rate": 0.0010132285218226855,
+ "loss": 3.8981378078460693,
+ "step": 2235
+ },
+ {
+ "epoch": 31.05589519650655,
+ "grad_norm": 0.009963045828044415,
+ "learning_rate": 0.0010119741884357424,
+ "loss": 3.881542444229126,
+ "step": 2236
+ },
+ {
+ "epoch": 31.069868995633186,
+ "grad_norm": 0.008536554872989655,
+ "learning_rate": 0.0010107202603276096,
+ "loss": 3.851253032684326,
+ "step": 2237
+ },
+ {
+ "epoch": 31.083842794759825,
+ "grad_norm": 0.01074561569839716,
+ "learning_rate": 0.001009466738508555,
+ "loss": 3.8731184005737305,
+ "step": 2238
+ },
+ {
+ "epoch": 31.097816593886463,
+ "grad_norm": 0.010809902101755142,
+ "learning_rate": 0.0010082136239885177,
+ "loss": 3.863889217376709,
+ "step": 2239
+ },
+ {
+ "epoch": 31.111790393013102,
+ "grad_norm": 0.008289352990686893,
+ "learning_rate": 0.0010069609177771103,
+ "loss": 3.839066743850708,
+ "step": 2240
+ },
+ {
+ "epoch": 31.125764192139737,
+ "grad_norm": 0.00792208593338728,
+ "learning_rate": 0.001005708620883615,
+ "loss": 3.8634843826293945,
+ "step": 2241
+ },
+ {
+ "epoch": 31.139737991266376,
+ "grad_norm": 0.008381598629057407,
+ "learning_rate": 0.0010044567343169845,
+ "loss": 3.868283987045288,
+ "step": 2242
+ },
+ {
+ "epoch": 31.153711790393015,
+ "grad_norm": 0.009784617461264133,
+ "learning_rate": 0.0010032052590858425,
+ "loss": 3.853074550628662,
+ "step": 2243
+ },
+ {
+ "epoch": 31.16768558951965,
+ "grad_norm": 0.009462198242545128,
+ "learning_rate": 0.0010019541961984782,
+ "loss": 3.889979600906372,
+ "step": 2244
+ },
+ {
+ "epoch": 31.18165938864629,
+ "grad_norm": 0.010207819752395153,
+ "learning_rate": 0.001000703546662852,
+ "loss": 3.8636741638183594,
+ "step": 2245
+ },
+ {
+ "epoch": 31.195633187772927,
+ "grad_norm": 0.009123551659286022,
+ "learning_rate": 0.0009994533114865887,
+ "loss": 3.8578555583953857,
+ "step": 2246
+ },
+ {
+ "epoch": 31.209606986899562,
+ "grad_norm": 0.009006140753626823,
+ "learning_rate": 0.0009982034916769806,
+ "loss": 3.8828015327453613,
+ "step": 2247
+ },
+ {
+ "epoch": 31.2235807860262,
+ "grad_norm": 0.0091114342212677,
+ "learning_rate": 0.0009969540882409846,
+ "loss": 3.880486011505127,
+ "step": 2248
+ },
+ {
+ "epoch": 31.23755458515284,
+ "grad_norm": 0.008851367980241776,
+ "learning_rate": 0.0009957051021852227,
+ "loss": 3.8618125915527344,
+ "step": 2249
+ },
+ {
+ "epoch": 31.251528384279474,
+ "grad_norm": 0.008196741342544556,
+ "learning_rate": 0.0009944565345159806,
+ "loss": 3.8314175605773926,
+ "step": 2250
+ },
+ {
+ "epoch": 31.265502183406113,
+ "grad_norm": 0.00863513071089983,
+ "learning_rate": 0.000993208386239206,
+ "loss": 3.856212615966797,
+ "step": 2251
+ },
+ {
+ "epoch": 31.27947598253275,
+ "grad_norm": 0.008624681271612644,
+ "learning_rate": 0.0009919606583605107,
+ "loss": 3.868927001953125,
+ "step": 2252
+ },
+ {
+ "epoch": 31.29344978165939,
+ "grad_norm": 0.009208842180669308,
+ "learning_rate": 0.0009907133518851647,
+ "loss": 3.9072189331054688,
+ "step": 2253
+ },
+ {
+ "epoch": 31.307423580786025,
+ "grad_norm": 0.010200913064181805,
+ "learning_rate": 0.0009894664678181027,
+ "loss": 3.875272750854492,
+ "step": 2254
+ },
+ {
+ "epoch": 31.321397379912664,
+ "grad_norm": 0.009892710484564304,
+ "learning_rate": 0.000988220007163914,
+ "loss": 3.888686418533325,
+ "step": 2255
+ },
+ {
+ "epoch": 31.335371179039303,
+ "grad_norm": 0.009873585775494576,
+ "learning_rate": 0.000986973970926851,
+ "loss": 3.8607118129730225,
+ "step": 2256
+ },
+ {
+ "epoch": 31.349344978165938,
+ "grad_norm": 0.009322221390902996,
+ "learning_rate": 0.0009857283601108228,
+ "loss": 3.8852052688598633,
+ "step": 2257
+ },
+ {
+ "epoch": 31.363318777292577,
+ "grad_norm": 0.008141528815031052,
+ "learning_rate": 0.0009844831757193947,
+ "loss": 3.888213872909546,
+ "step": 2258
+ },
+ {
+ "epoch": 31.377292576419215,
+ "grad_norm": 0.008191613480448723,
+ "learning_rate": 0.00098323841875579,
+ "loss": 3.9073967933654785,
+ "step": 2259
+ },
+ {
+ "epoch": 31.39126637554585,
+ "grad_norm": 0.008419081568717957,
+ "learning_rate": 0.0009819940902228864,
+ "loss": 3.879223346710205,
+ "step": 2260
+ },
+ {
+ "epoch": 31.40524017467249,
+ "grad_norm": 0.009037869982421398,
+ "learning_rate": 0.0009807501911232171,
+ "loss": 3.883303165435791,
+ "step": 2261
+ },
+ {
+ "epoch": 31.419213973799128,
+ "grad_norm": 0.00854078121483326,
+ "learning_rate": 0.0009795067224589687,
+ "loss": 3.926767587661743,
+ "step": 2262
+ },
+ {
+ "epoch": 31.433187772925763,
+ "grad_norm": 0.008590771816670895,
+ "learning_rate": 0.0009782636852319825,
+ "loss": 3.913999557495117,
+ "step": 2263
+ },
+ {
+ "epoch": 31.4471615720524,
+ "grad_norm": 0.00859815627336502,
+ "learning_rate": 0.0009770210804437498,
+ "loss": 3.8828063011169434,
+ "step": 2264
+ },
+ {
+ "epoch": 31.46113537117904,
+ "grad_norm": 0.008970260620117188,
+ "learning_rate": 0.0009757789090954157,
+ "loss": 3.9174866676330566,
+ "step": 2265
+ },
+ {
+ "epoch": 31.475109170305675,
+ "grad_norm": 0.00877175759524107,
+ "learning_rate": 0.0009745371721877754,
+ "loss": 3.8879401683807373,
+ "step": 2266
+ },
+ {
+ "epoch": 31.489082969432314,
+ "grad_norm": 0.008541764691472054,
+ "learning_rate": 0.0009732958707212731,
+ "loss": 3.8717427253723145,
+ "step": 2267
+ },
+ {
+ "epoch": 31.503056768558952,
+ "grad_norm": 0.008466241881251335,
+ "learning_rate": 0.0009720550056960039,
+ "loss": 3.8944485187530518,
+ "step": 2268
+ },
+ {
+ "epoch": 31.51703056768559,
+ "grad_norm": 0.00808491837233305,
+ "learning_rate": 0.0009708145781117096,
+ "loss": 3.8617026805877686,
+ "step": 2269
+ },
+ {
+ "epoch": 31.531004366812226,
+ "grad_norm": 0.00842207670211792,
+ "learning_rate": 0.0009695745889677805,
+ "loss": 3.887824535369873,
+ "step": 2270
+ },
+ {
+ "epoch": 31.544978165938865,
+ "grad_norm": 0.00830599945038557,
+ "learning_rate": 0.0009683350392632533,
+ "loss": 3.8612070083618164,
+ "step": 2271
+ },
+ {
+ "epoch": 31.558951965065503,
+ "grad_norm": 0.008417914621531963,
+ "learning_rate": 0.0009670959299968111,
+ "loss": 3.880140781402588,
+ "step": 2272
+ },
+ {
+ "epoch": 31.57292576419214,
+ "grad_norm": 0.008542059920728207,
+ "learning_rate": 0.0009658572621667822,
+ "loss": 3.8951029777526855,
+ "step": 2273
+ },
+ {
+ "epoch": 31.586899563318777,
+ "grad_norm": 0.008822236210107803,
+ "learning_rate": 0.000964619036771138,
+ "loss": 3.8962583541870117,
+ "step": 2274
+ },
+ {
+ "epoch": 31.600873362445416,
+ "grad_norm": 0.009151184931397438,
+ "learning_rate": 0.0009633812548074952,
+ "loss": 3.882500648498535,
+ "step": 2275
+ },
+ {
+ "epoch": 31.61484716157205,
+ "grad_norm": 0.007803036365658045,
+ "learning_rate": 0.0009621439172731115,
+ "loss": 3.8740782737731934,
+ "step": 2276
+ },
+ {
+ "epoch": 31.62882096069869,
+ "grad_norm": 0.009350850246846676,
+ "learning_rate": 0.0009609070251648884,
+ "loss": 3.8627419471740723,
+ "step": 2277
+ },
+ {
+ "epoch": 31.64279475982533,
+ "grad_norm": 0.008319144137203693,
+ "learning_rate": 0.0009596705794793667,
+ "loss": 3.9084486961364746,
+ "step": 2278
+ },
+ {
+ "epoch": 31.656768558951963,
+ "grad_norm": 0.009198528714478016,
+ "learning_rate": 0.0009584345812127282,
+ "loss": 3.92141056060791,
+ "step": 2279
+ },
+ {
+ "epoch": 31.670742358078602,
+ "grad_norm": 0.010480938479304314,
+ "learning_rate": 0.0009571990313607958,
+ "loss": 3.8916103839874268,
+ "step": 2280
+ },
+ {
+ "epoch": 31.68471615720524,
+ "grad_norm": 0.01045288611203432,
+ "learning_rate": 0.0009559639309190278,
+ "loss": 3.8863210678100586,
+ "step": 2281
+ },
+ {
+ "epoch": 31.69868995633188,
+ "grad_norm": 0.009121391922235489,
+ "learning_rate": 0.0009547292808825239,
+ "loss": 3.9167044162750244,
+ "step": 2282
+ },
+ {
+ "epoch": 31.712663755458514,
+ "grad_norm": 0.01093223039060831,
+ "learning_rate": 0.0009534950822460184,
+ "loss": 3.9046711921691895,
+ "step": 2283
+ },
+ {
+ "epoch": 31.726637554585153,
+ "grad_norm": 0.01214868575334549,
+ "learning_rate": 0.0009522613360038835,
+ "loss": 3.9126768112182617,
+ "step": 2284
+ },
+ {
+ "epoch": 31.74061135371179,
+ "grad_norm": 0.01055280864238739,
+ "learning_rate": 0.0009510280431501255,
+ "loss": 3.9214353561401367,
+ "step": 2285
+ },
+ {
+ "epoch": 31.754585152838427,
+ "grad_norm": 0.009021313861012459,
+ "learning_rate": 0.0009497952046783867,
+ "loss": 3.8795580863952637,
+ "step": 2286
+ },
+ {
+ "epoch": 31.768558951965066,
+ "grad_norm": 0.00863752793520689,
+ "learning_rate": 0.0009485628215819428,
+ "loss": 3.893232822418213,
+ "step": 2287
+ },
+ {
+ "epoch": 31.782532751091704,
+ "grad_norm": 0.009327626787126064,
+ "learning_rate": 0.0009473308948537021,
+ "loss": 3.917377471923828,
+ "step": 2288
+ },
+ {
+ "epoch": 31.79650655021834,
+ "grad_norm": 0.01003898587077856,
+ "learning_rate": 0.0009460994254862066,
+ "loss": 3.8818931579589844,
+ "step": 2289
+ },
+ {
+ "epoch": 31.810480349344978,
+ "grad_norm": 0.00892532430589199,
+ "learning_rate": 0.0009448684144716278,
+ "loss": 3.8761978149414062,
+ "step": 2290
+ },
+ {
+ "epoch": 31.824454148471617,
+ "grad_norm": 0.008396463468670845,
+ "learning_rate": 0.0009436378628017702,
+ "loss": 3.8665928840637207,
+ "step": 2291
+ },
+ {
+ "epoch": 31.83842794759825,
+ "grad_norm": 0.00784370768815279,
+ "learning_rate": 0.0009424077714680662,
+ "loss": 3.8850138187408447,
+ "step": 2292
+ },
+ {
+ "epoch": 31.85240174672489,
+ "grad_norm": 0.00809321366250515,
+ "learning_rate": 0.0009411781414615784,
+ "loss": 3.917078733444214,
+ "step": 2293
+ },
+ {
+ "epoch": 31.86637554585153,
+ "grad_norm": 0.008685288950800896,
+ "learning_rate": 0.000939948973772998,
+ "loss": 3.8865013122558594,
+ "step": 2294
+ },
+ {
+ "epoch": 31.880349344978168,
+ "grad_norm": 0.007291820365935564,
+ "learning_rate": 0.0009387202693926425,
+ "loss": 3.8967666625976562,
+ "step": 2295
+ },
+ {
+ "epoch": 31.894323144104803,
+ "grad_norm": 0.008459319360554218,
+ "learning_rate": 0.0009374920293104573,
+ "loss": 3.9043235778808594,
+ "step": 2296
+ },
+ {
+ "epoch": 31.90829694323144,
+ "grad_norm": 0.007924137637019157,
+ "learning_rate": 0.0009362642545160123,
+ "loss": 3.9216856956481934,
+ "step": 2297
+ },
+ {
+ "epoch": 31.92227074235808,
+ "grad_norm": 0.007874754257500172,
+ "learning_rate": 0.0009350369459985052,
+ "loss": 3.879725933074951,
+ "step": 2298
+ },
+ {
+ "epoch": 31.936244541484715,
+ "grad_norm": 0.008643723092973232,
+ "learning_rate": 0.0009338101047467544,
+ "loss": 3.9303886890411377,
+ "step": 2299
+ },
+ {
+ "epoch": 31.950218340611354,
+ "grad_norm": 0.008451704867184162,
+ "learning_rate": 0.000932583731749205,
+ "loss": 3.9352827072143555,
+ "step": 2300
+ },
+ {
+ "epoch": 31.964192139737992,
+ "grad_norm": 0.008137754164636135,
+ "learning_rate": 0.0009313578279939236,
+ "loss": 3.930800437927246,
+ "step": 2301
+ },
+ {
+ "epoch": 31.978165938864628,
+ "grad_norm": 0.009277567267417908,
+ "learning_rate": 0.0009301323944685979,
+ "loss": 3.9194300174713135,
+ "step": 2302
+ },
+ {
+ "epoch": 31.992139737991266,
+ "grad_norm": 0.009428970515727997,
+ "learning_rate": 0.0009289074321605383,
+ "loss": 3.874298572540283,
+ "step": 2303
+ },
+ {
+ "epoch": 32.0,
+ "grad_norm": 0.0089727733284235,
+ "learning_rate": 0.0009276829420566743,
+ "loss": 3.893770694732666,
+ "step": 2304
+ },
+ {
+ "epoch": 32.0,
+ "eval_loss": 4.388528823852539,
+ "eval_runtime": 51.5859,
+ "eval_samples_per_second": 47.338,
+ "eval_steps_per_second": 1.493,
+ "step": 2304
+ },
+ {
+ "epoch": 32.01397379912664,
+ "grad_norm": 0.009470812045037746,
+ "learning_rate": 0.000926458925143556,
+ "loss": 3.8514318466186523,
+ "step": 2305
+ },
+ {
+ "epoch": 32.02794759825328,
+ "grad_norm": 0.011485518887639046,
+ "learning_rate": 0.0009252353824073511,
+ "loss": 3.857734203338623,
+ "step": 2306
+ },
+ {
+ "epoch": 32.041921397379916,
+ "grad_norm": 0.010095939971506596,
+ "learning_rate": 0.0009240123148338457,
+ "loss": 3.819117546081543,
+ "step": 2307
+ },
+ {
+ "epoch": 32.05589519650655,
+ "grad_norm": 0.010103728622198105,
+ "learning_rate": 0.0009227897234084444,
+ "loss": 3.842660427093506,
+ "step": 2308
+ },
+ {
+ "epoch": 32.069868995633186,
+ "grad_norm": 0.01148147787898779,
+ "learning_rate": 0.0009215676091161661,
+ "loss": 3.821089029312134,
+ "step": 2309
+ },
+ {
+ "epoch": 32.083842794759825,
+ "grad_norm": 0.012331553734838963,
+ "learning_rate": 0.0009203459729416467,
+ "loss": 3.8194339275360107,
+ "step": 2310
+ },
+ {
+ "epoch": 32.09781659388646,
+ "grad_norm": 0.011274794116616249,
+ "learning_rate": 0.0009191248158691358,
+ "loss": 3.8839240074157715,
+ "step": 2311
+ },
+ {
+ "epoch": 32.1117903930131,
+ "grad_norm": 0.009603365324437618,
+ "learning_rate": 0.0009179041388824986,
+ "loss": 3.806617259979248,
+ "step": 2312
+ },
+ {
+ "epoch": 32.12576419213974,
+ "grad_norm": 0.010824795812368393,
+ "learning_rate": 0.0009166839429652112,
+ "loss": 3.823228597640991,
+ "step": 2313
+ },
+ {
+ "epoch": 32.13973799126637,
+ "grad_norm": 0.01014280878007412,
+ "learning_rate": 0.0009154642291003637,
+ "loss": 3.860605239868164,
+ "step": 2314
+ },
+ {
+ "epoch": 32.15371179039301,
+ "grad_norm": 0.009052124805748463,
+ "learning_rate": 0.0009142449982706591,
+ "loss": 3.84226131439209,
+ "step": 2315
+ },
+ {
+ "epoch": 32.16768558951965,
+ "grad_norm": 0.009465130977332592,
+ "learning_rate": 0.0009130262514584074,
+ "loss": 3.8674912452697754,
+ "step": 2316
+ },
+ {
+ "epoch": 32.18165938864629,
+ "grad_norm": 0.009991783648729324,
+ "learning_rate": 0.0009118079896455326,
+ "loss": 3.8775739669799805,
+ "step": 2317
+ },
+ {
+ "epoch": 32.19563318777293,
+ "grad_norm": 0.009288253262639046,
+ "learning_rate": 0.0009105902138135652,
+ "loss": 3.7978098392486572,
+ "step": 2318
+ },
+ {
+ "epoch": 32.209606986899566,
+ "grad_norm": 0.009964323602616787,
+ "learning_rate": 0.0009093729249436459,
+ "loss": 3.8362321853637695,
+ "step": 2319
+ },
+ {
+ "epoch": 32.223580786026204,
+ "grad_norm": 0.010977157391607761,
+ "learning_rate": 0.0009081561240165218,
+ "loss": 3.857304096221924,
+ "step": 2320
+ },
+ {
+ "epoch": 32.237554585152836,
+ "grad_norm": 0.009241319261491299,
+ "learning_rate": 0.0009069398120125479,
+ "loss": 3.8659310340881348,
+ "step": 2321
+ },
+ {
+ "epoch": 32.251528384279474,
+ "grad_norm": 0.009211038239300251,
+ "learning_rate": 0.0009057239899116844,
+ "loss": 3.8764538764953613,
+ "step": 2322
+ },
+ {
+ "epoch": 32.26550218340611,
+ "grad_norm": 0.009649050422012806,
+ "learning_rate": 0.000904508658693497,
+ "loss": 3.8702778816223145,
+ "step": 2323
+ },
+ {
+ "epoch": 32.27947598253275,
+ "grad_norm": 0.00954380165785551,
+ "learning_rate": 0.0009032938193371573,
+ "loss": 3.8273327350616455,
+ "step": 2324
+ },
+ {
+ "epoch": 32.29344978165939,
+ "grad_norm": 0.008831270970404148,
+ "learning_rate": 0.0009020794728214378,
+ "loss": 3.821664810180664,
+ "step": 2325
+ },
+ {
+ "epoch": 32.30742358078603,
+ "grad_norm": 0.008367286063730717,
+ "learning_rate": 0.000900865620124717,
+ "loss": 3.8586034774780273,
+ "step": 2326
+ },
+ {
+ "epoch": 32.32139737991266,
+ "grad_norm": 0.008805174380540848,
+ "learning_rate": 0.0008996522622249732,
+ "loss": 3.811522960662842,
+ "step": 2327
+ },
+ {
+ "epoch": 32.3353711790393,
+ "grad_norm": 0.00873226672410965,
+ "learning_rate": 0.0008984394000997873,
+ "loss": 3.8538150787353516,
+ "step": 2328
+ },
+ {
+ "epoch": 32.34934497816594,
+ "grad_norm": 0.008930203504860401,
+ "learning_rate": 0.0008972270347263402,
+ "loss": 3.8448486328125,
+ "step": 2329
+ },
+ {
+ "epoch": 32.36331877729258,
+ "grad_norm": 0.009300434030592442,
+ "learning_rate": 0.0008960151670814125,
+ "loss": 3.842113494873047,
+ "step": 2330
+ },
+ {
+ "epoch": 32.377292576419215,
+ "grad_norm": 0.009598663076758385,
+ "learning_rate": 0.0008948037981413855,
+ "loss": 3.886852741241455,
+ "step": 2331
+ },
+ {
+ "epoch": 32.391266375545854,
+ "grad_norm": 0.008400370366871357,
+ "learning_rate": 0.0008935929288822357,
+ "loss": 3.852076292037964,
+ "step": 2332
+ },
+ {
+ "epoch": 32.40524017467249,
+ "grad_norm": 0.00891344528645277,
+ "learning_rate": 0.0008923825602795391,
+ "loss": 3.8480606079101562,
+ "step": 2333
+ },
+ {
+ "epoch": 32.419213973799124,
+ "grad_norm": 0.00869396049529314,
+ "learning_rate": 0.0008911726933084678,
+ "loss": 3.8164706230163574,
+ "step": 2334
+ },
+ {
+ "epoch": 32.43318777292576,
+ "grad_norm": 0.007914266549050808,
+ "learning_rate": 0.0008899633289437904,
+ "loss": 3.8768272399902344,
+ "step": 2335
+ },
+ {
+ "epoch": 32.4471615720524,
+ "grad_norm": 0.008398533798754215,
+ "learning_rate": 0.0008887544681598687,
+ "loss": 3.8625502586364746,
+ "step": 2336
+ },
+ {
+ "epoch": 32.46113537117904,
+ "grad_norm": 0.00887146033346653,
+ "learning_rate": 0.0008875461119306601,
+ "loss": 3.8934593200683594,
+ "step": 2337
+ },
+ {
+ "epoch": 32.47510917030568,
+ "grad_norm": 0.008642833679914474,
+ "learning_rate": 0.0008863382612297172,
+ "loss": 3.8391976356506348,
+ "step": 2338
+ },
+ {
+ "epoch": 32.48908296943232,
+ "grad_norm": 0.008099788799881935,
+ "learning_rate": 0.0008851309170301816,
+ "loss": 3.827965497970581,
+ "step": 2339
+ },
+ {
+ "epoch": 32.50305676855895,
+ "grad_norm": 0.008166342042386532,
+ "learning_rate": 0.0008839240803047894,
+ "loss": 3.8999152183532715,
+ "step": 2340
+ },
+ {
+ "epoch": 32.51703056768559,
+ "grad_norm": 0.009095479734241962,
+ "learning_rate": 0.0008827177520258671,
+ "loss": 3.876927614212036,
+ "step": 2341
+ },
+ {
+ "epoch": 32.531004366812226,
+ "grad_norm": 0.009246554225683212,
+ "learning_rate": 0.0008815119331653321,
+ "loss": 3.8692450523376465,
+ "step": 2342
+ },
+ {
+ "epoch": 32.544978165938865,
+ "grad_norm": 0.007059223018586636,
+ "learning_rate": 0.0008803066246946898,
+ "loss": 3.869400978088379,
+ "step": 2343
+ },
+ {
+ "epoch": 32.5589519650655,
+ "grad_norm": 0.008499744348227978,
+ "learning_rate": 0.0008791018275850363,
+ "loss": 3.8303189277648926,
+ "step": 2344
+ },
+ {
+ "epoch": 32.57292576419214,
+ "grad_norm": 0.007667839527130127,
+ "learning_rate": 0.0008778975428070543,
+ "loss": 3.827998638153076,
+ "step": 2345
+ },
+ {
+ "epoch": 32.58689956331878,
+ "grad_norm": 0.00765089550986886,
+ "learning_rate": 0.0008766937713310145,
+ "loss": 3.8554482460021973,
+ "step": 2346
+ },
+ {
+ "epoch": 32.60087336244541,
+ "grad_norm": 0.00862938817590475,
+ "learning_rate": 0.0008754905141267749,
+ "loss": 3.905458927154541,
+ "step": 2347
+ },
+ {
+ "epoch": 32.61484716157205,
+ "grad_norm": 0.008733712136745453,
+ "learning_rate": 0.0008742877721637758,
+ "loss": 3.8564133644104004,
+ "step": 2348
+ },
+ {
+ "epoch": 32.62882096069869,
+ "grad_norm": 0.0071738120168447495,
+ "learning_rate": 0.0008730855464110471,
+ "loss": 3.878807544708252,
+ "step": 2349
+ },
+ {
+ "epoch": 32.64279475982533,
+ "grad_norm": 0.008418241515755653,
+ "learning_rate": 0.0008718838378371989,
+ "loss": 3.8399720191955566,
+ "step": 2350
+ },
+ {
+ "epoch": 32.65676855895197,
+ "grad_norm": 0.008526502177119255,
+ "learning_rate": 0.0008706826474104263,
+ "loss": 3.847656488418579,
+ "step": 2351
+ },
+ {
+ "epoch": 32.670742358078606,
+ "grad_norm": 0.00732462527230382,
+ "learning_rate": 0.0008694819760985069,
+ "loss": 3.8839492797851562,
+ "step": 2352
+ },
+ {
+ "epoch": 32.68471615720524,
+ "grad_norm": 0.008424196392297745,
+ "learning_rate": 0.0008682818248688003,
+ "loss": 3.861239433288574,
+ "step": 2353
+ },
+ {
+ "epoch": 32.698689956331876,
+ "grad_norm": 0.008984955959022045,
+ "learning_rate": 0.0008670821946882465,
+ "loss": 3.856529712677002,
+ "step": 2354
+ },
+ {
+ "epoch": 32.712663755458514,
+ "grad_norm": 0.009004034101963043,
+ "learning_rate": 0.000865883086523365,
+ "loss": 3.8514528274536133,
+ "step": 2355
+ },
+ {
+ "epoch": 32.72663755458515,
+ "grad_norm": 0.009097260423004627,
+ "learning_rate": 0.0008646845013402563,
+ "loss": 3.8890581130981445,
+ "step": 2356
+ },
+ {
+ "epoch": 32.74061135371179,
+ "grad_norm": 0.009244748391211033,
+ "learning_rate": 0.0008634864401045988,
+ "loss": 3.872745990753174,
+ "step": 2357
+ },
+ {
+ "epoch": 32.75458515283843,
+ "grad_norm": 0.00788815412670374,
+ "learning_rate": 0.0008622889037816488,
+ "loss": 3.8777575492858887,
+ "step": 2358
+ },
+ {
+ "epoch": 32.76855895196506,
+ "grad_norm": 0.008943610824644566,
+ "learning_rate": 0.0008610918933362396,
+ "loss": 3.8825736045837402,
+ "step": 2359
+ },
+ {
+ "epoch": 32.7825327510917,
+ "grad_norm": 0.008154108189046383,
+ "learning_rate": 0.0008598954097327809,
+ "loss": 3.877514362335205,
+ "step": 2360
+ },
+ {
+ "epoch": 32.79650655021834,
+ "grad_norm": 0.008689315989613533,
+ "learning_rate": 0.0008586994539352587,
+ "loss": 3.854720115661621,
+ "step": 2361
+ },
+ {
+ "epoch": 32.81048034934498,
+ "grad_norm": 0.00904505793005228,
+ "learning_rate": 0.0008575040269072315,
+ "loss": 3.911463499069214,
+ "step": 2362
+ },
+ {
+ "epoch": 32.82445414847162,
+ "grad_norm": 0.007773210760205984,
+ "learning_rate": 0.0008563091296118342,
+ "loss": 3.875370740890503,
+ "step": 2363
+ },
+ {
+ "epoch": 32.838427947598255,
+ "grad_norm": 0.008437697775661945,
+ "learning_rate": 0.0008551147630117737,
+ "loss": 3.8457469940185547,
+ "step": 2364
+ },
+ {
+ "epoch": 32.852401746724894,
+ "grad_norm": 0.009780957363545895,
+ "learning_rate": 0.0008539209280693296,
+ "loss": 3.890479564666748,
+ "step": 2365
+ },
+ {
+ "epoch": 32.866375545851525,
+ "grad_norm": 0.007929213345050812,
+ "learning_rate": 0.0008527276257463532,
+ "loss": 3.876030445098877,
+ "step": 2366
+ },
+ {
+ "epoch": 32.880349344978164,
+ "grad_norm": 0.009121386334300041,
+ "learning_rate": 0.0008515348570042669,
+ "loss": 3.9252734184265137,
+ "step": 2367
+ },
+ {
+ "epoch": 32.8943231441048,
+ "grad_norm": 0.00981531199067831,
+ "learning_rate": 0.000850342622804063,
+ "loss": 3.8997983932495117,
+ "step": 2368
+ },
+ {
+ "epoch": 32.90829694323144,
+ "grad_norm": 0.009150844067335129,
+ "learning_rate": 0.0008491509241063025,
+ "loss": 3.85115385055542,
+ "step": 2369
+ },
+ {
+ "epoch": 32.92227074235808,
+ "grad_norm": 0.00992978923022747,
+ "learning_rate": 0.0008479597618711154,
+ "loss": 3.9138102531433105,
+ "step": 2370
+ },
+ {
+ "epoch": 32.93624454148472,
+ "grad_norm": 0.010906689800322056,
+ "learning_rate": 0.0008467691370582001,
+ "loss": 3.892329692840576,
+ "step": 2371
+ },
+ {
+ "epoch": 32.95021834061135,
+ "grad_norm": 0.008644519373774529,
+ "learning_rate": 0.0008455790506268213,
+ "loss": 3.9071311950683594,
+ "step": 2372
+ },
+ {
+ "epoch": 32.96419213973799,
+ "grad_norm": 0.009030954912304878,
+ "learning_rate": 0.0008443895035358111,
+ "loss": 3.867384910583496,
+ "step": 2373
+ },
+ {
+ "epoch": 32.97816593886463,
+ "grad_norm": 0.010806960985064507,
+ "learning_rate": 0.0008432004967435638,
+ "loss": 3.877983331680298,
+ "step": 2374
+ },
+ {
+ "epoch": 32.992139737991266,
+ "grad_norm": 0.011148087680339813,
+ "learning_rate": 0.0008420120312080436,
+ "loss": 3.8795924186706543,
+ "step": 2375
+ },
+ {
+ "epoch": 33.0,
+ "grad_norm": 0.01046178862452507,
+ "learning_rate": 0.0008408241078867731,
+ "loss": 3.887709617614746,
+ "step": 2376
+ },
+ {
+ "epoch": 33.0,
+ "eval_loss": 4.383835792541504,
+ "eval_runtime": 53.0778,
+ "eval_samples_per_second": 46.008,
+ "eval_steps_per_second": 1.451,
+ "step": 2376
+ },
+ {
+ "epoch": 33.01397379912664,
+ "grad_norm": 0.009721867740154266,
+ "learning_rate": 0.0008396367277368421,
+ "loss": 3.8261327743530273,
+ "step": 2377
+ },
+ {
+ "epoch": 33.02794759825328,
+ "grad_norm": 0.01091420091688633,
+ "learning_rate": 0.000838449891714901,
+ "loss": 3.7837276458740234,
+ "step": 2378
+ },
+ {
+ "epoch": 33.041921397379916,
+ "grad_norm": 0.010423852130770683,
+ "learning_rate": 0.0008372636007771623,
+ "loss": 3.796450138092041,
+ "step": 2379
+ },
+ {
+ "epoch": 33.05589519650655,
+ "grad_norm": 0.011928882449865341,
+ "learning_rate": 0.0008360778558793996,
+ "loss": 3.794717311859131,
+ "step": 2380
+ },
+ {
+ "epoch": 33.069868995633186,
+ "grad_norm": 0.010779444128274918,
+ "learning_rate": 0.000834892657976945,
+ "loss": 3.8477983474731445,
+ "step": 2381
+ },
+ {
+ "epoch": 33.083842794759825,
+ "grad_norm": 0.009554282762110233,
+ "learning_rate": 0.0008337080080246915,
+ "loss": 3.793790817260742,
+ "step": 2382
+ },
+ {
+ "epoch": 33.09781659388646,
+ "grad_norm": 0.009768152609467506,
+ "learning_rate": 0.0008325239069770906,
+ "loss": 3.860384225845337,
+ "step": 2383
+ },
+ {
+ "epoch": 33.1117903930131,
+ "grad_norm": 0.010801832191646099,
+ "learning_rate": 0.0008313403557881513,
+ "loss": 3.816796064376831,
+ "step": 2384
+ },
+ {
+ "epoch": 33.12576419213974,
+ "grad_norm": 0.01046004518866539,
+ "learning_rate": 0.0008301573554114378,
+ "loss": 3.8064301013946533,
+ "step": 2385
+ },
+ {
+ "epoch": 33.13973799126637,
+ "grad_norm": 0.009319447912275791,
+ "learning_rate": 0.0008289749068000744,
+ "loss": 3.832408905029297,
+ "step": 2386
+ },
+ {
+ "epoch": 33.15371179039301,
+ "grad_norm": 0.01011407095938921,
+ "learning_rate": 0.000827793010906737,
+ "loss": 3.8276009559631348,
+ "step": 2387
+ },
+ {
+ "epoch": 33.16768558951965,
+ "grad_norm": 0.010393989272415638,
+ "learning_rate": 0.0008266116686836582,
+ "loss": 3.83754825592041,
+ "step": 2388
+ },
+ {
+ "epoch": 33.18165938864629,
+ "grad_norm": 0.00883004255592823,
+ "learning_rate": 0.0008254308810826244,
+ "loss": 3.8315513134002686,
+ "step": 2389
+ },
+ {
+ "epoch": 33.19563318777293,
+ "grad_norm": 0.008549787104129791,
+ "learning_rate": 0.0008242506490549745,
+ "loss": 3.792919158935547,
+ "step": 2390
+ },
+ {
+ "epoch": 33.209606986899566,
+ "grad_norm": 0.009668390266597271,
+ "learning_rate": 0.0008230709735516006,
+ "loss": 3.8110954761505127,
+ "step": 2391
+ },
+ {
+ "epoch": 33.223580786026204,
+ "grad_norm": 0.009008699096739292,
+ "learning_rate": 0.0008218918555229445,
+ "loss": 3.8598508834838867,
+ "step": 2392
+ },
+ {
+ "epoch": 33.237554585152836,
+ "grad_norm": 0.008660725317895412,
+ "learning_rate": 0.0008207132959190026,
+ "loss": 3.839632511138916,
+ "step": 2393
+ },
+ {
+ "epoch": 33.251528384279474,
+ "grad_norm": 0.00827972125262022,
+ "learning_rate": 0.0008195352956893172,
+ "loss": 3.8428683280944824,
+ "step": 2394
+ },
+ {
+ "epoch": 33.26550218340611,
+ "grad_norm": 0.008117233403027058,
+ "learning_rate": 0.0008183578557829825,
+ "loss": 3.8434956073760986,
+ "step": 2395
+ },
+ {
+ "epoch": 33.27947598253275,
+ "grad_norm": 0.008314455859363079,
+ "learning_rate": 0.0008171809771486405,
+ "loss": 3.8079686164855957,
+ "step": 2396
+ },
+ {
+ "epoch": 33.29344978165939,
+ "grad_norm": 0.00865850131958723,
+ "learning_rate": 0.0008160046607344811,
+ "loss": 3.846999406814575,
+ "step": 2397
+ },
+ {
+ "epoch": 33.30742358078603,
+ "grad_norm": 0.0096864253282547,
+ "learning_rate": 0.0008148289074882416,
+ "loss": 3.8227243423461914,
+ "step": 2398
+ },
+ {
+ "epoch": 33.32139737991266,
+ "grad_norm": 0.00840236246585846,
+ "learning_rate": 0.0008136537183572045,
+ "loss": 3.8400344848632812,
+ "step": 2399
+ },
+ {
+ "epoch": 33.3353711790393,
+ "grad_norm": 0.007451791316270828,
+ "learning_rate": 0.0008124790942881984,
+ "loss": 3.8318095207214355,
+ "step": 2400
+ },
+ {
+ "epoch": 33.34934497816594,
+ "grad_norm": 0.00858070608228445,
+ "learning_rate": 0.000811305036227597,
+ "loss": 3.8232922554016113,
+ "step": 2401
+ },
+ {
+ "epoch": 33.36331877729258,
+ "grad_norm": 0.00765061192214489,
+ "learning_rate": 0.0008101315451213178,
+ "loss": 3.81624174118042,
+ "step": 2402
+ },
+ {
+ "epoch": 33.377292576419215,
+ "grad_norm": 0.009304090403020382,
+ "learning_rate": 0.0008089586219148213,
+ "loss": 3.8715004920959473,
+ "step": 2403
+ },
+ {
+ "epoch": 33.391266375545854,
+ "grad_norm": 0.009136030450463295,
+ "learning_rate": 0.0008077862675531103,
+ "loss": 3.825037956237793,
+ "step": 2404
+ },
+ {
+ "epoch": 33.40524017467249,
+ "grad_norm": 0.007460939697921276,
+ "learning_rate": 0.0008066144829807303,
+ "loss": 3.811749219894409,
+ "step": 2405
+ },
+ {
+ "epoch": 33.419213973799124,
+ "grad_norm": 0.007510480936616659,
+ "learning_rate": 0.0008054432691417659,
+ "loss": 3.8313679695129395,
+ "step": 2406
+ },
+ {
+ "epoch": 33.43318777292576,
+ "grad_norm": 0.008825462311506271,
+ "learning_rate": 0.0008042726269798433,
+ "loss": 3.83119535446167,
+ "step": 2407
+ },
+ {
+ "epoch": 33.4471615720524,
+ "grad_norm": 0.009002318605780602,
+ "learning_rate": 0.0008031025574381281,
+ "loss": 3.8542468547821045,
+ "step": 2408
+ },
+ {
+ "epoch": 33.46113537117904,
+ "grad_norm": 0.009266122244298458,
+ "learning_rate": 0.0008019330614593237,
+ "loss": 3.8328065872192383,
+ "step": 2409
+ },
+ {
+ "epoch": 33.47510917030568,
+ "grad_norm": 0.008935115300118923,
+ "learning_rate": 0.0008007641399856727,
+ "loss": 3.8198719024658203,
+ "step": 2410
+ },
+ {
+ "epoch": 33.48908296943232,
+ "grad_norm": 0.008021850138902664,
+ "learning_rate": 0.0007995957939589526,
+ "loss": 3.828382968902588,
+ "step": 2411
+ },
+ {
+ "epoch": 33.50305676855895,
+ "grad_norm": 0.008035171777009964,
+ "learning_rate": 0.0007984280243204805,
+ "loss": 3.879868507385254,
+ "step": 2412
+ },
+ {
+ "epoch": 33.51703056768559,
+ "grad_norm": 0.008553579449653625,
+ "learning_rate": 0.0007972608320111059,
+ "loss": 3.842162609100342,
+ "step": 2413
+ },
+ {
+ "epoch": 33.531004366812226,
+ "grad_norm": 0.008016746491193771,
+ "learning_rate": 0.000796094217971215,
+ "loss": 3.85674786567688,
+ "step": 2414
+ },
+ {
+ "epoch": 33.544978165938865,
+ "grad_norm": 0.007781418971717358,
+ "learning_rate": 0.0007949281831407276,
+ "loss": 3.7968859672546387,
+ "step": 2415
+ },
+ {
+ "epoch": 33.5589519650655,
+ "grad_norm": 0.008552255108952522,
+ "learning_rate": 0.0007937627284590971,
+ "loss": 3.8305506706237793,
+ "step": 2416
+ },
+ {
+ "epoch": 33.57292576419214,
+ "grad_norm": 0.007390565704554319,
+ "learning_rate": 0.0007925978548653096,
+ "loss": 3.8228843212127686,
+ "step": 2417
+ },
+ {
+ "epoch": 33.58689956331878,
+ "grad_norm": 0.007611850742250681,
+ "learning_rate": 0.0007914335632978812,
+ "loss": 3.868957996368408,
+ "step": 2418
+ },
+ {
+ "epoch": 33.60087336244541,
+ "grad_norm": 0.008188549429178238,
+ "learning_rate": 0.0007902698546948628,
+ "loss": 3.836078405380249,
+ "step": 2419
+ },
+ {
+ "epoch": 33.61484716157205,
+ "grad_norm": 0.008198797702789307,
+ "learning_rate": 0.0007891067299938318,
+ "loss": 3.8637044429779053,
+ "step": 2420
+ },
+ {
+ "epoch": 33.62882096069869,
+ "grad_norm": 0.008353965356945992,
+ "learning_rate": 0.0007879441901318977,
+ "loss": 3.833843946456909,
+ "step": 2421
+ },
+ {
+ "epoch": 33.64279475982533,
+ "grad_norm": 0.0086967209354043,
+ "learning_rate": 0.0007867822360456963,
+ "loss": 3.850536823272705,
+ "step": 2422
+ },
+ {
+ "epoch": 33.65676855895197,
+ "grad_norm": 0.009151408448815346,
+ "learning_rate": 0.0007856208686713949,
+ "loss": 3.8453407287597656,
+ "step": 2423
+ },
+ {
+ "epoch": 33.670742358078606,
+ "grad_norm": 0.008010942488908768,
+ "learning_rate": 0.0007844600889446858,
+ "loss": 3.8679513931274414,
+ "step": 2424
+ },
+ {
+ "epoch": 33.68471615720524,
+ "grad_norm": 0.007813206873834133,
+ "learning_rate": 0.0007832998978007873,
+ "loss": 3.860609531402588,
+ "step": 2425
+ },
+ {
+ "epoch": 33.698689956331876,
+ "grad_norm": 0.007759213913232088,
+ "learning_rate": 0.0007821402961744448,
+ "loss": 3.803863048553467,
+ "step": 2426
+ },
+ {
+ "epoch": 33.712663755458514,
+ "grad_norm": 0.008045864291489124,
+ "learning_rate": 0.0007809812849999293,
+ "loss": 3.858161449432373,
+ "step": 2427
+ },
+ {
+ "epoch": 33.72663755458515,
+ "grad_norm": 0.008212368935346603,
+ "learning_rate": 0.0007798228652110347,
+ "loss": 3.8311336040496826,
+ "step": 2428
+ },
+ {
+ "epoch": 33.74061135371179,
+ "grad_norm": 0.007982932962477207,
+ "learning_rate": 0.0007786650377410774,
+ "loss": 3.870339870452881,
+ "step": 2429
+ },
+ {
+ "epoch": 33.75458515283843,
+ "grad_norm": 0.0071402136236429214,
+ "learning_rate": 0.0007775078035229002,
+ "loss": 3.7922263145446777,
+ "step": 2430
+ },
+ {
+ "epoch": 33.76855895196506,
+ "grad_norm": 0.008564750663936138,
+ "learning_rate": 0.0007763511634888656,
+ "loss": 3.83981990814209,
+ "step": 2431
+ },
+ {
+ "epoch": 33.7825327510917,
+ "grad_norm": 0.008078585378825665,
+ "learning_rate": 0.0007751951185708562,
+ "loss": 3.852938652038574,
+ "step": 2432
+ },
+ {
+ "epoch": 33.79650655021834,
+ "grad_norm": 0.00801754929125309,
+ "learning_rate": 0.0007740396697002774,
+ "loss": 3.8434009552001953,
+ "step": 2433
+ },
+ {
+ "epoch": 33.81048034934498,
+ "grad_norm": 0.008973612450063229,
+ "learning_rate": 0.0007728848178080535,
+ "loss": 3.8238563537597656,
+ "step": 2434
+ },
+ {
+ "epoch": 33.82445414847162,
+ "grad_norm": 0.010199865326285362,
+ "learning_rate": 0.0007717305638246281,
+ "loss": 3.8891735076904297,
+ "step": 2435
+ },
+ {
+ "epoch": 33.838427947598255,
+ "grad_norm": 0.009620471857488155,
+ "learning_rate": 0.0007705769086799617,
+ "loss": 3.8611769676208496,
+ "step": 2436
+ },
+ {
+ "epoch": 33.852401746724894,
+ "grad_norm": 0.007981563918292522,
+ "learning_rate": 0.0007694238533035337,
+ "loss": 3.8496928215026855,
+ "step": 2437
+ },
+ {
+ "epoch": 33.866375545851525,
+ "grad_norm": 0.008856567554175854,
+ "learning_rate": 0.0007682713986243415,
+ "loss": 3.8366098403930664,
+ "step": 2438
+ },
+ {
+ "epoch": 33.880349344978164,
+ "grad_norm": 0.008323253132402897,
+ "learning_rate": 0.0007671195455708952,
+ "loss": 3.8224244117736816,
+ "step": 2439
+ },
+ {
+ "epoch": 33.8943231441048,
+ "grad_norm": 0.00814893003553152,
+ "learning_rate": 0.0007659682950712228,
+ "loss": 3.8611154556274414,
+ "step": 2440
+ },
+ {
+ "epoch": 33.90829694323144,
+ "grad_norm": 0.009146672673523426,
+ "learning_rate": 0.0007648176480528659,
+ "loss": 3.8258843421936035,
+ "step": 2441
+ },
+ {
+ "epoch": 33.92227074235808,
+ "grad_norm": 0.007573423441499472,
+ "learning_rate": 0.0007636676054428805,
+ "loss": 3.888007640838623,
+ "step": 2442
+ },
+ {
+ "epoch": 33.93624454148472,
+ "grad_norm": 0.00813984964042902,
+ "learning_rate": 0.0007625181681678343,
+ "loss": 3.860306739807129,
+ "step": 2443
+ },
+ {
+ "epoch": 33.95021834061135,
+ "grad_norm": 0.007936778478324413,
+ "learning_rate": 0.0007613693371538078,
+ "loss": 3.8808555603027344,
+ "step": 2444
+ },
+ {
+ "epoch": 33.96419213973799,
+ "grad_norm": 0.008434669114649296,
+ "learning_rate": 0.0007602211133263956,
+ "loss": 3.854154348373413,
+ "step": 2445
+ },
+ {
+ "epoch": 33.97816593886463,
+ "grad_norm": 0.007857564836740494,
+ "learning_rate": 0.0007590734976106985,
+ "loss": 3.8240737915039062,
+ "step": 2446
+ },
+ {
+ "epoch": 33.992139737991266,
+ "grad_norm": 0.007989327423274517,
+ "learning_rate": 0.0007579264909313315,
+ "loss": 3.850971221923828,
+ "step": 2447
+ },
+ {
+ "epoch": 34.0,
+ "grad_norm": 0.008888393640518188,
+ "learning_rate": 0.0007567800942124152,
+ "loss": 3.862267255783081,
+ "step": 2448
+ },
+ {
+ "epoch": 34.0,
+ "eval_loss": 4.38022518157959,
+ "eval_runtime": 51.4686,
+ "eval_samples_per_second": 47.446,
+ "eval_steps_per_second": 1.496,
+ "step": 2448
+ },
+ {
+ "epoch": 34.01397379912664,
+ "grad_norm": 0.00952722318470478,
+ "learning_rate": 0.0007556343083775832,
+ "loss": 3.7705495357513428,
+ "step": 2449
+ },
+ {
+ "epoch": 34.02794759825328,
+ "grad_norm": 0.01099323108792305,
+ "learning_rate": 0.0007544891343499723,
+ "loss": 3.80812406539917,
+ "step": 2450
+ },
+ {
+ "epoch": 34.041921397379916,
+ "grad_norm": 0.010298571549355984,
+ "learning_rate": 0.0007533445730522295,
+ "loss": 3.782794237136841,
+ "step": 2451
+ },
+ {
+ "epoch": 34.05589519650655,
+ "grad_norm": 0.008139589801430702,
+ "learning_rate": 0.000752200625406507,
+ "loss": 3.7877392768859863,
+ "step": 2452
+ },
+ {
+ "epoch": 34.069868995633186,
+ "grad_norm": 0.009031252935528755,
+ "learning_rate": 0.0007510572923344628,
+ "loss": 3.8181564807891846,
+ "step": 2453
+ },
+ {
+ "epoch": 34.083842794759825,
+ "grad_norm": 0.00942459236830473,
+ "learning_rate": 0.0007499145747572603,
+ "loss": 3.807425022125244,
+ "step": 2454
+ },
+ {
+ "epoch": 34.09781659388646,
+ "grad_norm": 0.00922947097569704,
+ "learning_rate": 0.0007487724735955648,
+ "loss": 3.7923264503479004,
+ "step": 2455
+ },
+ {
+ "epoch": 34.1117903930131,
+ "grad_norm": 0.009551790542900562,
+ "learning_rate": 0.0007476309897695485,
+ "loss": 3.790083885192871,
+ "step": 2456
+ },
+ {
+ "epoch": 34.12576419213974,
+ "grad_norm": 0.008996761403977871,
+ "learning_rate": 0.0007464901241988831,
+ "loss": 3.836491107940674,
+ "step": 2457
+ },
+ {
+ "epoch": 34.13973799126637,
+ "grad_norm": 0.009766182862222195,
+ "learning_rate": 0.0007453498778027433,
+ "loss": 3.7829647064208984,
+ "step": 2458
+ },
+ {
+ "epoch": 34.15371179039301,
+ "grad_norm": 0.007708575110882521,
+ "learning_rate": 0.0007442102514998058,
+ "loss": 3.803400993347168,
+ "step": 2459
+ },
+ {
+ "epoch": 34.16768558951965,
+ "grad_norm": 0.008606902323663235,
+ "learning_rate": 0.0007430712462082464,
+ "loss": 3.810825824737549,
+ "step": 2460
+ },
+ {
+ "epoch": 34.18165938864629,
+ "grad_norm": 0.009462927468121052,
+ "learning_rate": 0.0007419328628457416,
+ "loss": 3.784748077392578,
+ "step": 2461
+ },
+ {
+ "epoch": 34.19563318777293,
+ "grad_norm": 0.010283366777002811,
+ "learning_rate": 0.0007407951023294655,
+ "loss": 3.744312286376953,
+ "step": 2462
+ },
+ {
+ "epoch": 34.209606986899566,
+ "grad_norm": 0.008952794596552849,
+ "learning_rate": 0.0007396579655760917,
+ "loss": 3.8098015785217285,
+ "step": 2463
+ },
+ {
+ "epoch": 34.223580786026204,
+ "grad_norm": 0.008446207270026207,
+ "learning_rate": 0.0007385214535017907,
+ "loss": 3.8306760787963867,
+ "step": 2464
+ },
+ {
+ "epoch": 34.237554585152836,
+ "grad_norm": 0.007988285273313522,
+ "learning_rate": 0.0007373855670222302,
+ "loss": 3.7976696491241455,
+ "step": 2465
+ },
+ {
+ "epoch": 34.251528384279474,
+ "grad_norm": 0.007586437277495861,
+ "learning_rate": 0.0007362503070525718,
+ "loss": 3.785891056060791,
+ "step": 2466
+ },
+ {
+ "epoch": 34.26550218340611,
+ "grad_norm": 0.00775605533272028,
+ "learning_rate": 0.0007351156745074757,
+ "loss": 3.8176379203796387,
+ "step": 2467
+ },
+ {
+ "epoch": 34.27947598253275,
+ "grad_norm": 0.007727691903710365,
+ "learning_rate": 0.0007339816703010948,
+ "loss": 3.791743278503418,
+ "step": 2468
+ },
+ {
+ "epoch": 34.29344978165939,
+ "grad_norm": 0.008311517536640167,
+ "learning_rate": 0.0007328482953470747,
+ "loss": 3.818769693374634,
+ "step": 2469
+ },
+ {
+ "epoch": 34.30742358078603,
+ "grad_norm": 0.007610874716192484,
+ "learning_rate": 0.0007317155505585556,
+ "loss": 3.76772141456604,
+ "step": 2470
+ },
+ {
+ "epoch": 34.32139737991266,
+ "grad_norm": 0.0078107016161084175,
+ "learning_rate": 0.0007305834368481699,
+ "loss": 3.822758197784424,
+ "step": 2471
+ },
+ {
+ "epoch": 34.3353711790393,
+ "grad_norm": 0.006906864698976278,
+ "learning_rate": 0.0007294519551280412,
+ "loss": 3.832834005355835,
+ "step": 2472
+ },
+ {
+ "epoch": 34.34934497816594,
+ "grad_norm": 0.008230660110712051,
+ "learning_rate": 0.0007283211063097829,
+ "loss": 3.85087513923645,
+ "step": 2473
+ },
+ {
+ "epoch": 34.36331877729258,
+ "grad_norm": 0.008201118558645248,
+ "learning_rate": 0.0007271908913044997,
+ "loss": 3.7777247428894043,
+ "step": 2474
+ },
+ {
+ "epoch": 34.377292576419215,
+ "grad_norm": 0.0070146857760846615,
+ "learning_rate": 0.000726061311022787,
+ "loss": 3.837517738342285,
+ "step": 2475
+ },
+ {
+ "epoch": 34.391266375545854,
+ "grad_norm": 0.0076036700047552586,
+ "learning_rate": 0.0007249323663747254,
+ "loss": 3.8358592987060547,
+ "step": 2476
+ },
+ {
+ "epoch": 34.40524017467249,
+ "grad_norm": 0.007238952908664942,
+ "learning_rate": 0.0007238040582698858,
+ "loss": 3.828089952468872,
+ "step": 2477
+ },
+ {
+ "epoch": 34.419213973799124,
+ "grad_norm": 0.00825269054621458,
+ "learning_rate": 0.0007226763876173259,
+ "loss": 3.806016445159912,
+ "step": 2478
+ },
+ {
+ "epoch": 34.43318777292576,
+ "grad_norm": 0.008368568494915962,
+ "learning_rate": 0.0007215493553255898,
+ "loss": 3.8163394927978516,
+ "step": 2479
+ },
+ {
+ "epoch": 34.4471615720524,
+ "grad_norm": 0.00728471577167511,
+ "learning_rate": 0.0007204229623027063,
+ "loss": 3.835597038269043,
+ "step": 2480
+ },
+ {
+ "epoch": 34.46113537117904,
+ "grad_norm": 0.00784830842167139,
+ "learning_rate": 0.0007192972094561897,
+ "loss": 3.8050971031188965,
+ "step": 2481
+ },
+ {
+ "epoch": 34.47510917030568,
+ "grad_norm": 0.007714141625910997,
+ "learning_rate": 0.0007181720976930404,
+ "loss": 3.8031606674194336,
+ "step": 2482
+ },
+ {
+ "epoch": 34.48908296943232,
+ "grad_norm": 0.009128103032708168,
+ "learning_rate": 0.000717047627919739,
+ "loss": 3.77107572555542,
+ "step": 2483
+ },
+ {
+ "epoch": 34.50305676855895,
+ "grad_norm": 0.008033272810280323,
+ "learning_rate": 0.000715923801042251,
+ "loss": 3.8615059852600098,
+ "step": 2484
+ },
+ {
+ "epoch": 34.51703056768559,
+ "grad_norm": 0.007732565049082041,
+ "learning_rate": 0.0007148006179660235,
+ "loss": 3.8277132511138916,
+ "step": 2485
+ },
+ {
+ "epoch": 34.531004366812226,
+ "grad_norm": 0.007382966112345457,
+ "learning_rate": 0.0007136780795959853,
+ "loss": 3.803431987762451,
+ "step": 2486
+ },
+ {
+ "epoch": 34.544978165938865,
+ "grad_norm": 0.007517020218074322,
+ "learning_rate": 0.0007125561868365442,
+ "loss": 3.799734354019165,
+ "step": 2487
+ },
+ {
+ "epoch": 34.5589519650655,
+ "grad_norm": 0.008041434921324253,
+ "learning_rate": 0.0007114349405915895,
+ "loss": 3.7848258018493652,
+ "step": 2488
+ },
+ {
+ "epoch": 34.57292576419214,
+ "grad_norm": 0.007875649258494377,
+ "learning_rate": 0.0007103143417644891,
+ "loss": 3.8136563301086426,
+ "step": 2489
+ },
+ {
+ "epoch": 34.58689956331878,
+ "grad_norm": 0.008213640190660954,
+ "learning_rate": 0.000709194391258089,
+ "loss": 3.8518495559692383,
+ "step": 2490
+ },
+ {
+ "epoch": 34.60087336244541,
+ "grad_norm": 0.00801054760813713,
+ "learning_rate": 0.0007080750899747139,
+ "loss": 3.7956271171569824,
+ "step": 2491
+ },
+ {
+ "epoch": 34.61484716157205,
+ "grad_norm": 0.008064007386565208,
+ "learning_rate": 0.0007069564388161631,
+ "loss": 3.854595184326172,
+ "step": 2492
+ },
+ {
+ "epoch": 34.62882096069869,
+ "grad_norm": 0.008300521411001682,
+ "learning_rate": 0.0007058384386837154,
+ "loss": 3.8203625679016113,
+ "step": 2493
+ },
+ {
+ "epoch": 34.64279475982533,
+ "grad_norm": 0.008640496991574764,
+ "learning_rate": 0.000704721090478122,
+ "loss": 3.793574810028076,
+ "step": 2494
+ },
+ {
+ "epoch": 34.65676855895197,
+ "grad_norm": 0.007684200536459684,
+ "learning_rate": 0.0007036043950996105,
+ "loss": 3.8564438819885254,
+ "step": 2495
+ },
+ {
+ "epoch": 34.670742358078606,
+ "grad_norm": 0.008431905880570412,
+ "learning_rate": 0.0007024883534478823,
+ "loss": 3.8271119594573975,
+ "step": 2496
+ },
+ {
+ "epoch": 34.68471615720524,
+ "grad_norm": 0.008137186989188194,
+ "learning_rate": 0.0007013729664221122,
+ "loss": 3.8219337463378906,
+ "step": 2497
+ },
+ {
+ "epoch": 34.698689956331876,
+ "grad_norm": 0.0076744635589420795,
+ "learning_rate": 0.0007002582349209474,
+ "loss": 3.8353049755096436,
+ "step": 2498
+ },
+ {
+ "epoch": 34.712663755458514,
+ "grad_norm": 0.008388740941882133,
+ "learning_rate": 0.0006991441598425065,
+ "loss": 3.795104503631592,
+ "step": 2499
+ },
+ {
+ "epoch": 34.72663755458515,
+ "grad_norm": 0.007856911979615688,
+ "learning_rate": 0.0006980307420843798,
+ "loss": 3.818411111831665,
+ "step": 2500
+ },
+ {
+ "epoch": 34.74061135371179,
+ "grad_norm": 0.008048887364566326,
+ "learning_rate": 0.000696917982543628,
+ "loss": 3.86599063873291,
+ "step": 2501
+ },
+ {
+ "epoch": 34.75458515283843,
+ "grad_norm": 0.007949979975819588,
+ "learning_rate": 0.0006958058821167812,
+ "loss": 3.8026742935180664,
+ "step": 2502
+ },
+ {
+ "epoch": 34.76855895196506,
+ "grad_norm": 0.008162821643054485,
+ "learning_rate": 0.0006946944416998389,
+ "loss": 3.833751678466797,
+ "step": 2503
+ },
+ {
+ "epoch": 34.7825327510917,
+ "grad_norm": 0.008190765976905823,
+ "learning_rate": 0.0006935836621882682,
+ "loss": 3.8164236545562744,
+ "step": 2504
+ },
+ {
+ "epoch": 34.79650655021834,
+ "grad_norm": 0.009714310988783836,
+ "learning_rate": 0.000692473544477005,
+ "loss": 3.8367366790771484,
+ "step": 2505
+ },
+ {
+ "epoch": 34.81048034934498,
+ "grad_norm": 0.008075987920165062,
+ "learning_rate": 0.0006913640894604498,
+ "loss": 3.8108773231506348,
+ "step": 2506
+ },
+ {
+ "epoch": 34.82445414847162,
+ "grad_norm": 0.009075167588889599,
+ "learning_rate": 0.0006902552980324713,
+ "loss": 3.8444485664367676,
+ "step": 2507
+ },
+ {
+ "epoch": 34.838427947598255,
+ "grad_norm": 0.008678440935909748,
+ "learning_rate": 0.0006891471710864022,
+ "loss": 3.8384461402893066,
+ "step": 2508
+ },
+ {
+ "epoch": 34.852401746724894,
+ "grad_norm": 0.007705411408096552,
+ "learning_rate": 0.0006880397095150414,
+ "loss": 3.825443744659424,
+ "step": 2509
+ },
+ {
+ "epoch": 34.866375545851525,
+ "grad_norm": 0.00782099086791277,
+ "learning_rate": 0.0006869329142106488,
+ "loss": 3.8331687450408936,
+ "step": 2510
+ },
+ {
+ "epoch": 34.880349344978164,
+ "grad_norm": 0.007542501203715801,
+ "learning_rate": 0.0006858267860649509,
+ "loss": 3.7905991077423096,
+ "step": 2511
+ },
+ {
+ "epoch": 34.8943231441048,
+ "grad_norm": 0.008315361104905605,
+ "learning_rate": 0.0006847213259691356,
+ "loss": 3.861527442932129,
+ "step": 2512
+ },
+ {
+ "epoch": 34.90829694323144,
+ "grad_norm": 0.008529976941645145,
+ "learning_rate": 0.0006836165348138508,
+ "loss": 3.8332972526550293,
+ "step": 2513
+ },
+ {
+ "epoch": 34.92227074235808,
+ "grad_norm": 0.008117211051285267,
+ "learning_rate": 0.0006825124134892072,
+ "loss": 3.7961089611053467,
+ "step": 2514
+ },
+ {
+ "epoch": 34.93624454148472,
+ "grad_norm": 0.008409497328102589,
+ "learning_rate": 0.0006814089628847755,
+ "loss": 3.812218189239502,
+ "step": 2515
+ },
+ {
+ "epoch": 34.95021834061135,
+ "grad_norm": 0.007592010777443647,
+ "learning_rate": 0.0006803061838895864,
+ "loss": 3.801304340362549,
+ "step": 2516
+ },
+ {
+ "epoch": 34.96419213973799,
+ "grad_norm": 0.007820053026080132,
+ "learning_rate": 0.0006792040773921283,
+ "loss": 3.8517117500305176,
+ "step": 2517
+ },
+ {
+ "epoch": 34.97816593886463,
+ "grad_norm": 0.0073044621385633945,
+ "learning_rate": 0.0006781026442803479,
+ "loss": 3.8253560066223145,
+ "step": 2518
+ },
+ {
+ "epoch": 34.992139737991266,
+ "grad_norm": 0.007546546868979931,
+ "learning_rate": 0.0006770018854416522,
+ "loss": 3.800802230834961,
+ "step": 2519
+ },
+ {
+ "epoch": 35.0,
+ "grad_norm": 0.008878474123775959,
+ "learning_rate": 0.0006759018017629007,
+ "loss": 3.8133859634399414,
+ "step": 2520
+ },
+ {
+ "epoch": 35.0,
+ "eval_loss": 4.375829219818115,
+ "eval_runtime": 51.3756,
+ "eval_samples_per_second": 47.532,
+ "eval_steps_per_second": 1.499,
+ "step": 2520
+ },
+ {
+ "epoch": 35.01397379912664,
+ "grad_norm": 0.00814585480839014,
+ "learning_rate": 0.0006748023941304115,
+ "loss": 3.7970919609069824,
+ "step": 2521
+ },
+ {
+ "epoch": 35.02794759825328,
+ "grad_norm": 0.00824262760579586,
+ "learning_rate": 0.0006737036634299577,
+ "loss": 3.750131607055664,
+ "step": 2522
+ },
+ {
+ "epoch": 35.041921397379916,
+ "grad_norm": 0.008675593882799149,
+ "learning_rate": 0.000672605610546767,
+ "loss": 3.791456699371338,
+ "step": 2523
+ },
+ {
+ "epoch": 35.05589519650655,
+ "grad_norm": 0.009052547626197338,
+ "learning_rate": 0.00067150823636552,
+ "loss": 3.7762060165405273,
+ "step": 2524
+ },
+ {
+ "epoch": 35.069868995633186,
+ "grad_norm": 0.00818016566336155,
+ "learning_rate": 0.0006704115417703514,
+ "loss": 3.750631332397461,
+ "step": 2525
+ },
+ {
+ "epoch": 35.083842794759825,
+ "grad_norm": 0.00800115056335926,
+ "learning_rate": 0.0006693155276448489,
+ "loss": 3.811519145965576,
+ "step": 2526
+ },
+ {
+ "epoch": 35.09781659388646,
+ "grad_norm": 0.009032433852553368,
+ "learning_rate": 0.0006682201948720509,
+ "loss": 3.7805609703063965,
+ "step": 2527
+ },
+ {
+ "epoch": 35.1117903930131,
+ "grad_norm": 0.0077886078506708145,
+ "learning_rate": 0.0006671255443344474,
+ "loss": 3.749366521835327,
+ "step": 2528
+ },
+ {
+ "epoch": 35.12576419213974,
+ "grad_norm": 0.008603204973042011,
+ "learning_rate": 0.0006660315769139774,
+ "loss": 3.7686312198638916,
+ "step": 2529
+ },
+ {
+ "epoch": 35.13973799126637,
+ "grad_norm": 0.008361323736608028,
+ "learning_rate": 0.0006649382934920329,
+ "loss": 3.740269660949707,
+ "step": 2530
+ },
+ {
+ "epoch": 35.15371179039301,
+ "grad_norm": 0.00948982872068882,
+ "learning_rate": 0.0006638456949494507,
+ "loss": 3.7874045372009277,
+ "step": 2531
+ },
+ {
+ "epoch": 35.16768558951965,
+ "grad_norm": 0.008958354592323303,
+ "learning_rate": 0.0006627537821665185,
+ "loss": 3.78987455368042,
+ "step": 2532
+ },
+ {
+ "epoch": 35.18165938864629,
+ "grad_norm": 0.008702470920979977,
+ "learning_rate": 0.0006616625560229704,
+ "loss": 3.751061201095581,
+ "step": 2533
+ },
+ {
+ "epoch": 35.19563318777293,
+ "grad_norm": 0.008424391970038414,
+ "learning_rate": 0.000660572017397988,
+ "loss": 3.80816650390625,
+ "step": 2534
+ },
+ {
+ "epoch": 35.209606986899566,
+ "grad_norm": 0.009022071026265621,
+ "learning_rate": 0.0006594821671701989,
+ "loss": 3.726067304611206,
+ "step": 2535
+ },
+ {
+ "epoch": 35.223580786026204,
+ "grad_norm": 0.007974586449563503,
+ "learning_rate": 0.0006583930062176739,
+ "loss": 3.78564453125,
+ "step": 2536
+ },
+ {
+ "epoch": 35.237554585152836,
+ "grad_norm": 0.007887622341513634,
+ "learning_rate": 0.0006573045354179329,
+ "loss": 3.7993059158325195,
+ "step": 2537
+ },
+ {
+ "epoch": 35.251528384279474,
+ "grad_norm": 0.007782666943967342,
+ "learning_rate": 0.0006562167556479352,
+ "loss": 3.7440507411956787,
+ "step": 2538
+ },
+ {
+ "epoch": 35.26550218340611,
+ "grad_norm": 0.007923880591988564,
+ "learning_rate": 0.000655129667784086,
+ "loss": 3.7920784950256348,
+ "step": 2539
+ },
+ {
+ "epoch": 35.27947598253275,
+ "grad_norm": 0.007850701920688152,
+ "learning_rate": 0.0006540432727022325,
+ "loss": 3.7833633422851562,
+ "step": 2540
+ },
+ {
+ "epoch": 35.29344978165939,
+ "grad_norm": 0.007855751551687717,
+ "learning_rate": 0.0006529575712776633,
+ "loss": 3.812349319458008,
+ "step": 2541
+ },
+ {
+ "epoch": 35.30742358078603,
+ "grad_norm": 0.007832377217710018,
+ "learning_rate": 0.0006518725643851091,
+ "loss": 3.7815146446228027,
+ "step": 2542
+ },
+ {
+ "epoch": 35.32139737991266,
+ "grad_norm": 0.006997724529355764,
+ "learning_rate": 0.0006507882528987395,
+ "loss": 3.795346975326538,
+ "step": 2543
+ },
+ {
+ "epoch": 35.3353711790393,
+ "grad_norm": 0.007433670572936535,
+ "learning_rate": 0.0006497046376921648,
+ "loss": 3.841017246246338,
+ "step": 2544
+ },
+ {
+ "epoch": 35.34934497816594,
+ "grad_norm": 0.0075767990201711655,
+ "learning_rate": 0.0006486217196384344,
+ "loss": 3.7913427352905273,
+ "step": 2545
+ },
+ {
+ "epoch": 35.36331877729258,
+ "grad_norm": 0.007551983464509249,
+ "learning_rate": 0.0006475394996100359,
+ "loss": 3.765105962753296,
+ "step": 2546
+ },
+ {
+ "epoch": 35.377292576419215,
+ "grad_norm": 0.007235546596348286,
+ "learning_rate": 0.0006464579784788943,
+ "loss": 3.7668676376342773,
+ "step": 2547
+ },
+ {
+ "epoch": 35.391266375545854,
+ "grad_norm": 0.007872921414673328,
+ "learning_rate": 0.0006453771571163716,
+ "loss": 3.770536422729492,
+ "step": 2548
+ },
+ {
+ "epoch": 35.40524017467249,
+ "grad_norm": 0.007598782889544964,
+ "learning_rate": 0.0006442970363932667,
+ "loss": 3.796818733215332,
+ "step": 2549
+ },
+ {
+ "epoch": 35.419213973799124,
+ "grad_norm": 0.007902889512479305,
+ "learning_rate": 0.0006432176171798124,
+ "loss": 3.824531078338623,
+ "step": 2550
+ },
+ {
+ "epoch": 35.43318777292576,
+ "grad_norm": 0.008964902721345425,
+ "learning_rate": 0.0006421389003456776,
+ "loss": 3.781778335571289,
+ "step": 2551
+ },
+ {
+ "epoch": 35.4471615720524,
+ "grad_norm": 0.009936625137925148,
+ "learning_rate": 0.000641060886759965,
+ "loss": 3.750248670578003,
+ "step": 2552
+ },
+ {
+ "epoch": 35.46113537117904,
+ "grad_norm": 0.008381214924156666,
+ "learning_rate": 0.0006399835772912106,
+ "loss": 3.797954797744751,
+ "step": 2553
+ },
+ {
+ "epoch": 35.47510917030568,
+ "grad_norm": 0.007752922363579273,
+ "learning_rate": 0.000638906972807384,
+ "loss": 3.789350748062134,
+ "step": 2554
+ },
+ {
+ "epoch": 35.48908296943232,
+ "grad_norm": 0.009501751512289047,
+ "learning_rate": 0.0006378310741758841,
+ "loss": 3.791079521179199,
+ "step": 2555
+ },
+ {
+ "epoch": 35.50305676855895,
+ "grad_norm": 0.00936005637049675,
+ "learning_rate": 0.0006367558822635453,
+ "loss": 3.78501033782959,
+ "step": 2556
+ },
+ {
+ "epoch": 35.51703056768559,
+ "grad_norm": 0.008370982483029366,
+ "learning_rate": 0.0006356813979366284,
+ "loss": 3.792647361755371,
+ "step": 2557
+ },
+ {
+ "epoch": 35.531004366812226,
+ "grad_norm": 0.00782831758260727,
+ "learning_rate": 0.000634607622060827,
+ "loss": 3.772569179534912,
+ "step": 2558
+ },
+ {
+ "epoch": 35.544978165938865,
+ "grad_norm": 0.00830658059567213,
+ "learning_rate": 0.0006335345555012627,
+ "loss": 3.7787656784057617,
+ "step": 2559
+ },
+ {
+ "epoch": 35.5589519650655,
+ "grad_norm": 0.009015833027660847,
+ "learning_rate": 0.000632462199122486,
+ "loss": 3.787418842315674,
+ "step": 2560
+ },
+ {
+ "epoch": 35.57292576419214,
+ "grad_norm": 0.008993860334157944,
+ "learning_rate": 0.0006313905537884752,
+ "loss": 3.818718433380127,
+ "step": 2561
+ },
+ {
+ "epoch": 35.58689956331878,
+ "grad_norm": 0.007954884320497513,
+ "learning_rate": 0.0006303196203626344,
+ "loss": 3.8100008964538574,
+ "step": 2562
+ },
+ {
+ "epoch": 35.60087336244541,
+ "grad_norm": 0.007742907386273146,
+ "learning_rate": 0.0006292493997077973,
+ "loss": 3.7786598205566406,
+ "step": 2563
+ },
+ {
+ "epoch": 35.61484716157205,
+ "grad_norm": 0.008748230524361134,
+ "learning_rate": 0.0006281798926862195,
+ "loss": 3.798635482788086,
+ "step": 2564
+ },
+ {
+ "epoch": 35.62882096069869,
+ "grad_norm": 0.008306438103318214,
+ "learning_rate": 0.000627111100159585,
+ "loss": 3.827213764190674,
+ "step": 2565
+ },
+ {
+ "epoch": 35.64279475982533,
+ "grad_norm": 0.008027350530028343,
+ "learning_rate": 0.0006260430229889987,
+ "loss": 3.7682337760925293,
+ "step": 2566
+ },
+ {
+ "epoch": 35.65676855895197,
+ "grad_norm": 0.007779776118695736,
+ "learning_rate": 0.0006249756620349931,
+ "loss": 3.7515475749969482,
+ "step": 2567
+ },
+ {
+ "epoch": 35.670742358078606,
+ "grad_norm": 0.008748390711843967,
+ "learning_rate": 0.000623909018157521,
+ "loss": 3.7963085174560547,
+ "step": 2568
+ },
+ {
+ "epoch": 35.68471615720524,
+ "grad_norm": 0.008587505668401718,
+ "learning_rate": 0.0006228430922159577,
+ "loss": 3.7767529487609863,
+ "step": 2569
+ },
+ {
+ "epoch": 35.698689956331876,
+ "grad_norm": 0.007172323763370514,
+ "learning_rate": 0.0006217778850691007,
+ "loss": 3.79768705368042,
+ "step": 2570
+ },
+ {
+ "epoch": 35.712663755458514,
+ "grad_norm": 0.007224083412438631,
+ "learning_rate": 0.000620713397575168,
+ "loss": 3.785390853881836,
+ "step": 2571
+ },
+ {
+ "epoch": 35.72663755458515,
+ "grad_norm": 0.008209837600588799,
+ "learning_rate": 0.0006196496305917987,
+ "loss": 3.809828281402588,
+ "step": 2572
+ },
+ {
+ "epoch": 35.74061135371179,
+ "grad_norm": 0.007760563865303993,
+ "learning_rate": 0.0006185865849760491,
+ "loss": 3.838716506958008,
+ "step": 2573
+ },
+ {
+ "epoch": 35.75458515283843,
+ "grad_norm": 0.007362260017544031,
+ "learning_rate": 0.0006175242615843977,
+ "loss": 3.7840137481689453,
+ "step": 2574
+ },
+ {
+ "epoch": 35.76855895196506,
+ "grad_norm": 0.008719335310161114,
+ "learning_rate": 0.0006164626612727381,
+ "loss": 3.8228893280029297,
+ "step": 2575
+ },
+ {
+ "epoch": 35.7825327510917,
+ "grad_norm": 0.00856029987335205,
+ "learning_rate": 0.0006154017848963826,
+ "loss": 3.80448317527771,
+ "step": 2576
+ },
+ {
+ "epoch": 35.79650655021834,
+ "grad_norm": 0.008078506216406822,
+ "learning_rate": 0.0006143416333100606,
+ "loss": 3.7881453037261963,
+ "step": 2577
+ },
+ {
+ "epoch": 35.81048034934498,
+ "grad_norm": 0.007957885973155499,
+ "learning_rate": 0.0006132822073679164,
+ "loss": 3.814870834350586,
+ "step": 2578
+ },
+ {
+ "epoch": 35.82445414847162,
+ "grad_norm": 0.009362339973449707,
+ "learning_rate": 0.0006122235079235114,
+ "loss": 3.798335075378418,
+ "step": 2579
+ },
+ {
+ "epoch": 35.838427947598255,
+ "grad_norm": 0.008272632025182247,
+ "learning_rate": 0.0006111655358298194,
+ "loss": 3.8212976455688477,
+ "step": 2580
+ },
+ {
+ "epoch": 35.852401746724894,
+ "grad_norm": 0.007548908703029156,
+ "learning_rate": 0.0006101082919392301,
+ "loss": 3.8312830924987793,
+ "step": 2581
+ },
+ {
+ "epoch": 35.866375545851525,
+ "grad_norm": 0.007543765474110842,
+ "learning_rate": 0.0006090517771035457,
+ "loss": 3.8302183151245117,
+ "step": 2582
+ },
+ {
+ "epoch": 35.880349344978164,
+ "grad_norm": 0.007520351093262434,
+ "learning_rate": 0.000607995992173981,
+ "loss": 3.7898802757263184,
+ "step": 2583
+ },
+ {
+ "epoch": 35.8943231441048,
+ "grad_norm": 0.007168850861489773,
+ "learning_rate": 0.000606940938001163,
+ "loss": 3.8078577518463135,
+ "step": 2584
+ },
+ {
+ "epoch": 35.90829694323144,
+ "grad_norm": 0.007420001085847616,
+ "learning_rate": 0.0006058866154351296,
+ "loss": 3.799571990966797,
+ "step": 2585
+ },
+ {
+ "epoch": 35.92227074235808,
+ "grad_norm": 0.007621012162417173,
+ "learning_rate": 0.0006048330253253303,
+ "loss": 3.833843946456909,
+ "step": 2586
+ },
+ {
+ "epoch": 35.93624454148472,
+ "grad_norm": 0.007661828305572271,
+ "learning_rate": 0.0006037801685206222,
+ "loss": 3.820556640625,
+ "step": 2587
+ },
+ {
+ "epoch": 35.95021834061135,
+ "grad_norm": 0.007657836657017469,
+ "learning_rate": 0.0006027280458692738,
+ "loss": 3.798994541168213,
+ "step": 2588
+ },
+ {
+ "epoch": 35.96419213973799,
+ "grad_norm": 0.007722760550677776,
+ "learning_rate": 0.0006016766582189613,
+ "loss": 3.8589251041412354,
+ "step": 2589
+ },
+ {
+ "epoch": 35.97816593886463,
+ "grad_norm": 0.00789247453212738,
+ "learning_rate": 0.0006006260064167685,
+ "loss": 3.8673787117004395,
+ "step": 2590
+ },
+ {
+ "epoch": 35.992139737991266,
+ "grad_norm": 0.009137817658483982,
+ "learning_rate": 0.0005995760913091874,
+ "loss": 3.801938056945801,
+ "step": 2591
+ },
+ {
+ "epoch": 36.0,
+ "grad_norm": 0.008374894969165325,
+ "learning_rate": 0.0005985269137421141,
+ "loss": 3.820045232772827,
+ "step": 2592
+ },
+ {
+ "epoch": 36.0,
+ "eval_loss": 4.3730878829956055,
+ "eval_runtime": 48.1607,
+ "eval_samples_per_second": 50.705,
+ "eval_steps_per_second": 1.599,
+ "step": 2592
+ },
+ {
+ "epoch": 36.01397379912664,
+ "grad_norm": 0.009492472745478153,
+ "learning_rate": 0.0005974784745608536,
+ "loss": 3.738617420196533,
+ "step": 2593
+ },
+ {
+ "epoch": 36.02794759825328,
+ "grad_norm": 0.010017318651080132,
+ "learning_rate": 0.0005964307746101135,
+ "loss": 3.7551231384277344,
+ "step": 2594
+ },
+ {
+ "epoch": 36.041921397379916,
+ "grad_norm": 0.008958659134805202,
+ "learning_rate": 0.000595383814734007,
+ "loss": 3.7739763259887695,
+ "step": 2595
+ },
+ {
+ "epoch": 36.05589519650655,
+ "grad_norm": 0.008034365251660347,
+ "learning_rate": 0.0005943375957760505,
+ "loss": 3.7678627967834473,
+ "step": 2596
+ },
+ {
+ "epoch": 36.069868995633186,
+ "grad_norm": 0.008489780128002167,
+ "learning_rate": 0.000593292118579164,
+ "loss": 3.720778465270996,
+ "step": 2597
+ },
+ {
+ "epoch": 36.083842794759825,
+ "grad_norm": 0.009107839316129684,
+ "learning_rate": 0.0005922473839856699,
+ "loss": 3.7224440574645996,
+ "step": 2598
+ },
+ {
+ "epoch": 36.09781659388646,
+ "grad_norm": 0.009295578114688396,
+ "learning_rate": 0.0005912033928372906,
+ "loss": 3.7695257663726807,
+ "step": 2599
+ },
+ {
+ "epoch": 36.1117903930131,
+ "grad_norm": 0.007896862924098969,
+ "learning_rate": 0.0005901601459751532,
+ "loss": 3.742567300796509,
+ "step": 2600
+ },
+ {
+ "epoch": 36.12576419213974,
+ "grad_norm": 0.008227395825088024,
+ "learning_rate": 0.0005891176442397808,
+ "loss": 3.7547202110290527,
+ "step": 2601
+ },
+ {
+ "epoch": 36.13973799126637,
+ "grad_norm": 0.009326796047389507,
+ "learning_rate": 0.0005880758884710992,
+ "loss": 3.724808931350708,
+ "step": 2602
+ },
+ {
+ "epoch": 36.15371179039301,
+ "grad_norm": 0.007786506786942482,
+ "learning_rate": 0.0005870348795084319,
+ "loss": 3.7403640747070312,
+ "step": 2603
+ },
+ {
+ "epoch": 36.16768558951965,
+ "grad_norm": 0.00821041688323021,
+ "learning_rate": 0.0005859946181905009,
+ "loss": 3.770012617111206,
+ "step": 2604
+ },
+ {
+ "epoch": 36.18165938864629,
+ "grad_norm": 0.009544595144689083,
+ "learning_rate": 0.000584955105355427,
+ "loss": 3.7378525733947754,
+ "step": 2605
+ },
+ {
+ "epoch": 36.19563318777293,
+ "grad_norm": 0.008973932825028896,
+ "learning_rate": 0.0005839163418407254,
+ "loss": 3.7184295654296875,
+ "step": 2606
+ },
+ {
+ "epoch": 36.209606986899566,
+ "grad_norm": 0.009023688733577728,
+ "learning_rate": 0.00058287832848331,
+ "loss": 3.7388229370117188,
+ "step": 2607
+ },
+ {
+ "epoch": 36.223580786026204,
+ "grad_norm": 0.008151066489517689,
+ "learning_rate": 0.0005818410661194889,
+ "loss": 3.7632875442504883,
+ "step": 2608
+ },
+ {
+ "epoch": 36.237554585152836,
+ "grad_norm": 0.008495030924677849,
+ "learning_rate": 0.0005808045555849668,
+ "loss": 3.8114736080169678,
+ "step": 2609
+ },
+ {
+ "epoch": 36.251528384279474,
+ "grad_norm": 0.008739210665225983,
+ "learning_rate": 0.0005797687977148394,
+ "loss": 3.7647829055786133,
+ "step": 2610
+ },
+ {
+ "epoch": 36.26550218340611,
+ "grad_norm": 0.008062886074185371,
+ "learning_rate": 0.0005787337933435999,
+ "loss": 3.782172441482544,
+ "step": 2611
+ },
+ {
+ "epoch": 36.27947598253275,
+ "grad_norm": 0.007615398149937391,
+ "learning_rate": 0.0005776995433051326,
+ "loss": 3.7564642429351807,
+ "step": 2612
+ },
+ {
+ "epoch": 36.29344978165939,
+ "grad_norm": 0.007938290014863014,
+ "learning_rate": 0.0005766660484327133,
+ "loss": 3.756208896636963,
+ "step": 2613
+ },
+ {
+ "epoch": 36.30742358078603,
+ "grad_norm": 0.008517666719853878,
+ "learning_rate": 0.0005756333095590102,
+ "loss": 3.76729416847229,
+ "step": 2614
+ },
+ {
+ "epoch": 36.32139737991266,
+ "grad_norm": 0.007426336407661438,
+ "learning_rate": 0.0005746013275160827,
+ "loss": 3.74001407623291,
+ "step": 2615
+ },
+ {
+ "epoch": 36.3353711790393,
+ "grad_norm": 0.007664753124117851,
+ "learning_rate": 0.0005735701031353804,
+ "loss": 3.7819876670837402,
+ "step": 2616
+ },
+ {
+ "epoch": 36.34934497816594,
+ "grad_norm": 0.007903067395091057,
+ "learning_rate": 0.0005725396372477414,
+ "loss": 3.781219959259033,
+ "step": 2617
+ },
+ {
+ "epoch": 36.36331877729258,
+ "grad_norm": 0.007251975126564503,
+ "learning_rate": 0.0005715099306833931,
+ "loss": 3.763620615005493,
+ "step": 2618
+ },
+ {
+ "epoch": 36.377292576419215,
+ "grad_norm": 0.007422176189720631,
+ "learning_rate": 0.0005704809842719533,
+ "loss": 3.775099754333496,
+ "step": 2619
+ },
+ {
+ "epoch": 36.391266375545854,
+ "grad_norm": 0.009161600843071938,
+ "learning_rate": 0.0005694527988424239,
+ "loss": 3.792268991470337,
+ "step": 2620
+ },
+ {
+ "epoch": 36.40524017467249,
+ "grad_norm": 0.008330529555678368,
+ "learning_rate": 0.0005684253752231958,
+ "loss": 3.7967824935913086,
+ "step": 2621
+ },
+ {
+ "epoch": 36.419213973799124,
+ "grad_norm": 0.0073860278353095055,
+ "learning_rate": 0.0005673987142420458,
+ "loss": 3.76505970954895,
+ "step": 2622
+ },
+ {
+ "epoch": 36.43318777292576,
+ "grad_norm": 0.008462096564471722,
+ "learning_rate": 0.0005663728167261363,
+ "loss": 3.740450143814087,
+ "step": 2623
+ },
+ {
+ "epoch": 36.4471615720524,
+ "grad_norm": 0.00838128849864006,
+ "learning_rate": 0.0005653476835020138,
+ "loss": 3.7863235473632812,
+ "step": 2624
+ },
+ {
+ "epoch": 36.46113537117904,
+ "grad_norm": 0.007381014991551638,
+ "learning_rate": 0.0005643233153956092,
+ "loss": 3.772179126739502,
+ "step": 2625
+ },
+ {
+ "epoch": 36.47510917030568,
+ "grad_norm": 0.007363761775195599,
+ "learning_rate": 0.0005632997132322392,
+ "loss": 3.7793867588043213,
+ "step": 2626
+ },
+ {
+ "epoch": 36.48908296943232,
+ "grad_norm": 0.008684969507157803,
+ "learning_rate": 0.0005622768778366001,
+ "loss": 3.7736339569091797,
+ "step": 2627
+ },
+ {
+ "epoch": 36.50305676855895,
+ "grad_norm": 0.007934059016406536,
+ "learning_rate": 0.0005612548100327722,
+ "loss": 3.8040053844451904,
+ "step": 2628
+ },
+ {
+ "epoch": 36.51703056768559,
+ "grad_norm": 0.007136524189263582,
+ "learning_rate": 0.0005602335106442172,
+ "loss": 3.7589468955993652,
+ "step": 2629
+ },
+ {
+ "epoch": 36.531004366812226,
+ "grad_norm": 0.00789916142821312,
+ "learning_rate": 0.0005592129804937779,
+ "loss": 3.7953195571899414,
+ "step": 2630
+ },
+ {
+ "epoch": 36.544978165938865,
+ "grad_norm": 0.007919179275631905,
+ "learning_rate": 0.0005581932204036761,
+ "loss": 3.7539429664611816,
+ "step": 2631
+ },
+ {
+ "epoch": 36.5589519650655,
+ "grad_norm": 0.007734223268926144,
+ "learning_rate": 0.0005571742311955147,
+ "loss": 3.7620325088500977,
+ "step": 2632
+ },
+ {
+ "epoch": 36.57292576419214,
+ "grad_norm": 0.007005508057773113,
+ "learning_rate": 0.0005561560136902749,
+ "loss": 3.762538433074951,
+ "step": 2633
+ },
+ {
+ "epoch": 36.58689956331878,
+ "grad_norm": 0.008137146942317486,
+ "learning_rate": 0.0005551385687083161,
+ "loss": 3.773789405822754,
+ "step": 2634
+ },
+ {
+ "epoch": 36.60087336244541,
+ "grad_norm": 0.007310070563107729,
+ "learning_rate": 0.0005541218970693757,
+ "loss": 3.8077168464660645,
+ "step": 2635
+ },
+ {
+ "epoch": 36.61484716157205,
+ "grad_norm": 0.007335536181926727,
+ "learning_rate": 0.0005531059995925661,
+ "loss": 3.7761921882629395,
+ "step": 2636
+ },
+ {
+ "epoch": 36.62882096069869,
+ "grad_norm": 0.008022893220186234,
+ "learning_rate": 0.00055209087709638,
+ "loss": 3.7769880294799805,
+ "step": 2637
+ },
+ {
+ "epoch": 36.64279475982533,
+ "grad_norm": 0.008027128875255585,
+ "learning_rate": 0.0005510765303986814,
+ "loss": 3.7709951400756836,
+ "step": 2638
+ },
+ {
+ "epoch": 36.65676855895197,
+ "grad_norm": 0.00697955721989274,
+ "learning_rate": 0.0005500629603167118,
+ "loss": 3.7855210304260254,
+ "step": 2639
+ },
+ {
+ "epoch": 36.670742358078606,
+ "grad_norm": 0.007455077022314072,
+ "learning_rate": 0.0005490501676670861,
+ "loss": 3.7850632667541504,
+ "step": 2640
+ },
+ {
+ "epoch": 36.68471615720524,
+ "grad_norm": 0.0069657545536756516,
+ "learning_rate": 0.0005480381532657932,
+ "loss": 3.742671251296997,
+ "step": 2641
+ },
+ {
+ "epoch": 36.698689956331876,
+ "grad_norm": 0.007609451189637184,
+ "learning_rate": 0.0005470269179281953,
+ "loss": 3.793771266937256,
+ "step": 2642
+ },
+ {
+ "epoch": 36.712663755458514,
+ "grad_norm": 0.007437967229634523,
+ "learning_rate": 0.0005460164624690255,
+ "loss": 3.778846263885498,
+ "step": 2643
+ },
+ {
+ "epoch": 36.72663755458515,
+ "grad_norm": 0.007123028859496117,
+ "learning_rate": 0.00054500678770239,
+ "loss": 3.768375873565674,
+ "step": 2644
+ },
+ {
+ "epoch": 36.74061135371179,
+ "grad_norm": 0.008176218718290329,
+ "learning_rate": 0.0005439978944417654,
+ "loss": 3.82069730758667,
+ "step": 2645
+ },
+ {
+ "epoch": 36.75458515283843,
+ "grad_norm": 0.006943001411855221,
+ "learning_rate": 0.000542989783499999,
+ "loss": 3.7731552124023438,
+ "step": 2646
+ },
+ {
+ "epoch": 36.76855895196506,
+ "grad_norm": 0.0072205523028969765,
+ "learning_rate": 0.0005419824556893069,
+ "loss": 3.7724058628082275,
+ "step": 2647
+ },
+ {
+ "epoch": 36.7825327510917,
+ "grad_norm": 0.007656400091946125,
+ "learning_rate": 0.0005409759118212758,
+ "loss": 3.780989646911621,
+ "step": 2648
+ },
+ {
+ "epoch": 36.79650655021834,
+ "grad_norm": 0.006972154602408409,
+ "learning_rate": 0.0005399701527068599,
+ "loss": 3.7971787452697754,
+ "step": 2649
+ },
+ {
+ "epoch": 36.81048034934498,
+ "grad_norm": 0.007674331311136484,
+ "learning_rate": 0.0005389651791563806,
+ "loss": 3.798673391342163,
+ "step": 2650
+ },
+ {
+ "epoch": 36.82445414847162,
+ "grad_norm": 0.006952045019716024,
+ "learning_rate": 0.0005379609919795268,
+ "loss": 3.7982821464538574,
+ "step": 2651
+ },
+ {
+ "epoch": 36.838427947598255,
+ "grad_norm": 0.007882075384259224,
+ "learning_rate": 0.0005369575919853545,
+ "loss": 3.825962781906128,
+ "step": 2652
+ },
+ {
+ "epoch": 36.852401746724894,
+ "grad_norm": 0.007143386639654636,
+ "learning_rate": 0.0005359549799822854,
+ "loss": 3.826598644256592,
+ "step": 2653
+ },
+ {
+ "epoch": 36.866375545851525,
+ "grad_norm": 0.007587207015603781,
+ "learning_rate": 0.0005349531567781041,
+ "loss": 3.80623197555542,
+ "step": 2654
+ },
+ {
+ "epoch": 36.880349344978164,
+ "grad_norm": 0.007677781395614147,
+ "learning_rate": 0.0005339521231799632,
+ "loss": 3.7987256050109863,
+ "step": 2655
+ },
+ {
+ "epoch": 36.8943231441048,
+ "grad_norm": 0.007372343447059393,
+ "learning_rate": 0.0005329518799943773,
+ "loss": 3.7733969688415527,
+ "step": 2656
+ },
+ {
+ "epoch": 36.90829694323144,
+ "grad_norm": 0.00861304346472025,
+ "learning_rate": 0.0005319524280272231,
+ "loss": 3.7903685569763184,
+ "step": 2657
+ },
+ {
+ "epoch": 36.92227074235808,
+ "grad_norm": 0.00816237274557352,
+ "learning_rate": 0.0005309537680837418,
+ "loss": 3.77058744430542,
+ "step": 2658
+ },
+ {
+ "epoch": 36.93624454148472,
+ "grad_norm": 0.007624274119734764,
+ "learning_rate": 0.0005299559009685356,
+ "loss": 3.787982940673828,
+ "step": 2659
+ },
+ {
+ "epoch": 36.95021834061135,
+ "grad_norm": 0.007416521664708853,
+ "learning_rate": 0.0005289588274855683,
+ "loss": 3.7991762161254883,
+ "step": 2660
+ },
+ {
+ "epoch": 36.96419213973799,
+ "grad_norm": 0.007461614441126585,
+ "learning_rate": 0.0005279625484381632,
+ "loss": 3.8036863803863525,
+ "step": 2661
+ },
+ {
+ "epoch": 36.97816593886463,
+ "grad_norm": 0.007766335736960173,
+ "learning_rate": 0.0005269670646290043,
+ "loss": 3.809243679046631,
+ "step": 2662
+ },
+ {
+ "epoch": 36.992139737991266,
+ "grad_norm": 0.0073394919745624065,
+ "learning_rate": 0.0005259723768601365,
+ "loss": 3.756606101989746,
+ "step": 2663
+ },
+ {
+ "epoch": 37.0,
+ "grad_norm": 0.008771199733018875,
+ "learning_rate": 0.0005249784859329602,
+ "loss": 3.773951768875122,
+ "step": 2664
+ },
+ {
+ "epoch": 37.0,
+ "eval_loss": 4.370637893676758,
+ "eval_runtime": 49.4904,
+ "eval_samples_per_second": 49.343,
+ "eval_steps_per_second": 1.556,
+ "step": 2664
+ },
+ {
+ "epoch": 37.01397379912664,
+ "grad_norm": 0.008164486847817898,
+ "learning_rate": 0.0005239853926482358,
+ "loss": 3.7619595527648926,
+ "step": 2665
+ },
+ {
+ "epoch": 37.02794759825328,
+ "grad_norm": 0.007980132475495338,
+ "learning_rate": 0.0005229930978060806,
+ "loss": 3.7398204803466797,
+ "step": 2666
+ },
+ {
+ "epoch": 37.041921397379916,
+ "grad_norm": 0.0077056242153048515,
+ "learning_rate": 0.0005220016022059692,
+ "loss": 3.699580669403076,
+ "step": 2667
+ },
+ {
+ "epoch": 37.05589519650655,
+ "grad_norm": 0.0074698119424283504,
+ "learning_rate": 0.0005210109066467308,
+ "loss": 3.7299928665161133,
+ "step": 2668
+ },
+ {
+ "epoch": 37.069868995633186,
+ "grad_norm": 0.008384468965232372,
+ "learning_rate": 0.0005200210119265512,
+ "loss": 3.704989433288574,
+ "step": 2669
+ },
+ {
+ "epoch": 37.083842794759825,
+ "grad_norm": 0.008066141977906227,
+ "learning_rate": 0.0005190319188429709,
+ "loss": 3.751697540283203,
+ "step": 2670
+ },
+ {
+ "epoch": 37.09781659388646,
+ "grad_norm": 0.007213297300040722,
+ "learning_rate": 0.0005180436281928841,
+ "loss": 3.740692377090454,
+ "step": 2671
+ },
+ {
+ "epoch": 37.1117903930131,
+ "grad_norm": 0.007343464996665716,
+ "learning_rate": 0.0005170561407725396,
+ "loss": 3.7060189247131348,
+ "step": 2672
+ },
+ {
+ "epoch": 37.12576419213974,
+ "grad_norm": 0.00783725269138813,
+ "learning_rate": 0.0005160694573775363,
+ "loss": 3.7414426803588867,
+ "step": 2673
+ },
+ {
+ "epoch": 37.13973799126637,
+ "grad_norm": 0.007640232797712088,
+ "learning_rate": 0.0005150835788028294,
+ "loss": 3.7451486587524414,
+ "step": 2674
+ },
+ {
+ "epoch": 37.15371179039301,
+ "grad_norm": 0.008129071444272995,
+ "learning_rate": 0.0005140985058427218,
+ "loss": 3.734839916229248,
+ "step": 2675
+ },
+ {
+ "epoch": 37.16768558951965,
+ "grad_norm": 0.006920623127371073,
+ "learning_rate": 0.0005131142392908697,
+ "loss": 3.7138614654541016,
+ "step": 2676
+ },
+ {
+ "epoch": 37.18165938864629,
+ "grad_norm": 0.00798740889877081,
+ "learning_rate": 0.0005121307799402787,
+ "loss": 3.76496958732605,
+ "step": 2677
+ },
+ {
+ "epoch": 37.19563318777293,
+ "grad_norm": 0.006689135450869799,
+ "learning_rate": 0.0005111481285833045,
+ "loss": 3.7707948684692383,
+ "step": 2678
+ },
+ {
+ "epoch": 37.209606986899566,
+ "grad_norm": 0.007544311694800854,
+ "learning_rate": 0.0005101662860116516,
+ "loss": 3.7393479347229004,
+ "step": 2679
+ },
+ {
+ "epoch": 37.223580786026204,
+ "grad_norm": 0.007487910334020853,
+ "learning_rate": 0.0005091852530163717,
+ "loss": 3.7486305236816406,
+ "step": 2680
+ },
+ {
+ "epoch": 37.237554585152836,
+ "grad_norm": 0.00772349676117301,
+ "learning_rate": 0.0005082050303878673,
+ "loss": 3.753629207611084,
+ "step": 2681
+ },
+ {
+ "epoch": 37.251528384279474,
+ "grad_norm": 0.007883663289248943,
+ "learning_rate": 0.0005072256189158844,
+ "loss": 3.7497611045837402,
+ "step": 2682
+ },
+ {
+ "epoch": 37.26550218340611,
+ "grad_norm": 0.008153906092047691,
+ "learning_rate": 0.0005062470193895178,
+ "loss": 3.7459521293640137,
+ "step": 2683
+ },
+ {
+ "epoch": 37.27947598253275,
+ "grad_norm": 0.007083641365170479,
+ "learning_rate": 0.0005052692325972075,
+ "loss": 3.725599765777588,
+ "step": 2684
+ },
+ {
+ "epoch": 37.29344978165939,
+ "grad_norm": 0.007102683652192354,
+ "learning_rate": 0.0005042922593267385,
+ "loss": 3.7724428176879883,
+ "step": 2685
+ },
+ {
+ "epoch": 37.30742358078603,
+ "grad_norm": 0.008104706183075905,
+ "learning_rate": 0.0005033161003652409,
+ "loss": 3.7386724948883057,
+ "step": 2686
+ },
+ {
+ "epoch": 37.32139737991266,
+ "grad_norm": 0.0073765418492257595,
+ "learning_rate": 0.0005023407564991878,
+ "loss": 3.7478132247924805,
+ "step": 2687
+ },
+ {
+ "epoch": 37.3353711790393,
+ "grad_norm": 0.007110628765076399,
+ "learning_rate": 0.0005013662285143963,
+ "loss": 3.771925687789917,
+ "step": 2688
+ },
+ {
+ "epoch": 37.34934497816594,
+ "grad_norm": 0.007323848083615303,
+ "learning_rate": 0.000500392517196026,
+ "loss": 3.7460904121398926,
+ "step": 2689
+ },
+ {
+ "epoch": 37.36331877729258,
+ "grad_norm": 0.007830807007849216,
+ "learning_rate": 0.0004994196233285786,
+ "loss": 3.751222610473633,
+ "step": 2690
+ },
+ {
+ "epoch": 37.377292576419215,
+ "grad_norm": 0.007710007485002279,
+ "learning_rate": 0.0004984475476958973,
+ "loss": 3.739858627319336,
+ "step": 2691
+ },
+ {
+ "epoch": 37.391266375545854,
+ "grad_norm": 0.007293973118066788,
+ "learning_rate": 0.0004974762910811655,
+ "loss": 3.7410120964050293,
+ "step": 2692
+ },
+ {
+ "epoch": 37.40524017467249,
+ "grad_norm": 0.0073824538849294186,
+ "learning_rate": 0.0004965058542669078,
+ "loss": 3.711397647857666,
+ "step": 2693
+ },
+ {
+ "epoch": 37.419213973799124,
+ "grad_norm": 0.007156423293054104,
+ "learning_rate": 0.0004955362380349866,
+ "loss": 3.7881593704223633,
+ "step": 2694
+ },
+ {
+ "epoch": 37.43318777292576,
+ "grad_norm": 0.007705818396061659,
+ "learning_rate": 0.0004945674431666047,
+ "loss": 3.7639269828796387,
+ "step": 2695
+ },
+ {
+ "epoch": 37.4471615720524,
+ "grad_norm": 0.0075300210155546665,
+ "learning_rate": 0.0004935994704423026,
+ "loss": 3.7562999725341797,
+ "step": 2696
+ },
+ {
+ "epoch": 37.46113537117904,
+ "grad_norm": 0.007137116510421038,
+ "learning_rate": 0.000492632320641959,
+ "loss": 3.7659149169921875,
+ "step": 2697
+ },
+ {
+ "epoch": 37.47510917030568,
+ "grad_norm": 0.007225630804896355,
+ "learning_rate": 0.0004916659945447878,
+ "loss": 3.7625155448913574,
+ "step": 2698
+ },
+ {
+ "epoch": 37.48908296943232,
+ "grad_norm": 0.008027992211282253,
+ "learning_rate": 0.0004907004929293408,
+ "loss": 3.746478796005249,
+ "step": 2699
+ },
+ {
+ "epoch": 37.50305676855895,
+ "grad_norm": 0.007582477759569883,
+ "learning_rate": 0.0004897358165735066,
+ "loss": 3.759408950805664,
+ "step": 2700
+ },
+ {
+ "epoch": 37.51703056768559,
+ "grad_norm": 0.007575645577162504,
+ "learning_rate": 0.000488771966254506,
+ "loss": 3.7430644035339355,
+ "step": 2701
+ },
+ {
+ "epoch": 37.531004366812226,
+ "grad_norm": 0.008404700085520744,
+ "learning_rate": 0.0004878089427488963,
+ "loss": 3.7727198600769043,
+ "step": 2702
+ },
+ {
+ "epoch": 37.544978165938865,
+ "grad_norm": 0.007118664216250181,
+ "learning_rate": 0.00048684674683256815,
+ "loss": 3.747460126876831,
+ "step": 2703
+ },
+ {
+ "epoch": 37.5589519650655,
+ "grad_norm": 0.007220778148621321,
+ "learning_rate": 0.00048588537928074584,
+ "loss": 3.746413230895996,
+ "step": 2704
+ },
+ {
+ "epoch": 37.57292576419214,
+ "grad_norm": 0.0076194945722818375,
+ "learning_rate": 0.0004849248408679849,
+ "loss": 3.7808125019073486,
+ "step": 2705
+ },
+ {
+ "epoch": 37.58689956331878,
+ "grad_norm": 0.007285538129508495,
+ "learning_rate": 0.00048396513236817387,
+ "loss": 3.7750301361083984,
+ "step": 2706
+ },
+ {
+ "epoch": 37.60087336244541,
+ "grad_norm": 0.007698678877204657,
+ "learning_rate": 0.00048300625455453383,
+ "loss": 3.7474594116210938,
+ "step": 2707
+ },
+ {
+ "epoch": 37.61484716157205,
+ "grad_norm": 0.007687424309551716,
+ "learning_rate": 0.0004820482081996141,
+ "loss": 3.7929720878601074,
+ "step": 2708
+ },
+ {
+ "epoch": 37.62882096069869,
+ "grad_norm": 0.007363605313003063,
+ "learning_rate": 0.0004810909940752959,
+ "loss": 3.75911021232605,
+ "step": 2709
+ },
+ {
+ "epoch": 37.64279475982533,
+ "grad_norm": 0.007791243493556976,
+ "learning_rate": 0.00048013461295278836,
+ "loss": 3.7567291259765625,
+ "step": 2710
+ },
+ {
+ "epoch": 37.65676855895197,
+ "grad_norm": 0.007375436834990978,
+ "learning_rate": 0.0004791790656026323,
+ "loss": 3.7784414291381836,
+ "step": 2711
+ },
+ {
+ "epoch": 37.670742358078606,
+ "grad_norm": 0.007330593653023243,
+ "learning_rate": 0.0004782243527946934,
+ "loss": 3.7533810138702393,
+ "step": 2712
+ },
+ {
+ "epoch": 37.68471615720524,
+ "grad_norm": 0.006926248781383038,
+ "learning_rate": 0.00047727047529816694,
+ "loss": 3.7795910835266113,
+ "step": 2713
+ },
+ {
+ "epoch": 37.698689956331876,
+ "grad_norm": 0.007102136500179768,
+ "learning_rate": 0.00047631743388157496,
+ "loss": 3.812549114227295,
+ "step": 2714
+ },
+ {
+ "epoch": 37.712663755458514,
+ "grad_norm": 0.006676106248050928,
+ "learning_rate": 0.00047536522931276555,
+ "loss": 3.7578792572021484,
+ "step": 2715
+ },
+ {
+ "epoch": 37.72663755458515,
+ "grad_norm": 0.006926767062395811,
+ "learning_rate": 0.0004744138623589134,
+ "loss": 3.782416820526123,
+ "step": 2716
+ },
+ {
+ "epoch": 37.74061135371179,
+ "grad_norm": 0.006894777063280344,
+ "learning_rate": 0.000473463333786516,
+ "loss": 3.7840042114257812,
+ "step": 2717
+ },
+ {
+ "epoch": 37.75458515283843,
+ "grad_norm": 0.007393381092697382,
+ "learning_rate": 0.00047251364436139923,
+ "loss": 3.7532780170440674,
+ "step": 2718
+ },
+ {
+ "epoch": 37.76855895196506,
+ "grad_norm": 0.006510896608233452,
+ "learning_rate": 0.0004715647948487092,
+ "loss": 3.7618460655212402,
+ "step": 2719
+ },
+ {
+ "epoch": 37.7825327510917,
+ "grad_norm": 0.007283828221261501,
+ "learning_rate": 0.0004706167860129171,
+ "loss": 3.768498420715332,
+ "step": 2720
+ },
+ {
+ "epoch": 37.79650655021834,
+ "grad_norm": 0.00655644154176116,
+ "learning_rate": 0.0004696696186178168,
+ "loss": 3.7724053859710693,
+ "step": 2721
+ },
+ {
+ "epoch": 37.81048034934498,
+ "grad_norm": 0.007338619790971279,
+ "learning_rate": 0.0004687232934265239,
+ "loss": 3.7814595699310303,
+ "step": 2722
+ },
+ {
+ "epoch": 37.82445414847162,
+ "grad_norm": 0.006839970126748085,
+ "learning_rate": 0.0004677778112014758,
+ "loss": 3.771808624267578,
+ "step": 2723
+ },
+ {
+ "epoch": 37.838427947598255,
+ "grad_norm": 0.006776005961000919,
+ "learning_rate": 0.0004668331727044297,
+ "loss": 3.786924123764038,
+ "step": 2724
+ },
+ {
+ "epoch": 37.852401746724894,
+ "grad_norm": 0.007011357229202986,
+ "learning_rate": 0.00046588937869646415,
+ "loss": 3.787672996520996,
+ "step": 2725
+ },
+ {
+ "epoch": 37.866375545851525,
+ "grad_norm": 0.006482086144387722,
+ "learning_rate": 0.0004649464299379771,
+ "loss": 3.7794556617736816,
+ "step": 2726
+ },
+ {
+ "epoch": 37.880349344978164,
+ "grad_norm": 0.0069654518738389015,
+ "learning_rate": 0.000464004327188685,
+ "loss": 3.7520360946655273,
+ "step": 2727
+ },
+ {
+ "epoch": 37.8943231441048,
+ "grad_norm": 0.006526473443955183,
+ "learning_rate": 0.0004630630712076235,
+ "loss": 3.734144926071167,
+ "step": 2728
+ },
+ {
+ "epoch": 37.90829694323144,
+ "grad_norm": 0.006822322495281696,
+ "learning_rate": 0.00046212266275314504,
+ "loss": 3.7614707946777344,
+ "step": 2729
+ },
+ {
+ "epoch": 37.92227074235808,
+ "grad_norm": 0.006827171426266432,
+ "learning_rate": 0.00046118310258292045,
+ "loss": 3.755218029022217,
+ "step": 2730
+ },
+ {
+ "epoch": 37.93624454148472,
+ "grad_norm": 0.00706718722358346,
+ "learning_rate": 0.00046024439145393546,
+ "loss": 3.7397823333740234,
+ "step": 2731
+ },
+ {
+ "epoch": 37.95021834061135,
+ "grad_norm": 0.00689381267875433,
+ "learning_rate": 0.00045930653012249314,
+ "loss": 3.758922576904297,
+ "step": 2732
+ },
+ {
+ "epoch": 37.96419213973799,
+ "grad_norm": 0.007069882471114397,
+ "learning_rate": 0.0004583695193442113,
+ "loss": 3.759767532348633,
+ "step": 2733
+ },
+ {
+ "epoch": 37.97816593886463,
+ "grad_norm": 0.006889182608574629,
+ "learning_rate": 0.0004574333598740228,
+ "loss": 3.7555460929870605,
+ "step": 2734
+ },
+ {
+ "epoch": 37.992139737991266,
+ "grad_norm": 0.006866258103400469,
+ "learning_rate": 0.00045649805246617475,
+ "loss": 3.762068271636963,
+ "step": 2735
+ },
+ {
+ "epoch": 38.0,
+ "grad_norm": 0.008696220815181732,
+ "learning_rate": 0.000455563597874226,
+ "loss": 3.7110466957092285,
+ "step": 2736
+ },
+ {
+ "epoch": 38.0,
+ "eval_loss": 4.369833469390869,
+ "eval_runtime": 48.7852,
+ "eval_samples_per_second": 50.056,
+ "eval_steps_per_second": 1.578,
+ "step": 2736
+ },
+ {
+ "epoch": 38.01397379912664,
+ "grad_norm": 0.0077184769324958324,
+ "learning_rate": 0.0004546299968510516,
+ "loss": 3.704669713973999,
+ "step": 2737
+ },
+ {
+ "epoch": 38.02794759825328,
+ "grad_norm": 0.00846019759774208,
+ "learning_rate": 0.0004536972501488358,
+ "loss": 3.701132297515869,
+ "step": 2738
+ },
+ {
+ "epoch": 38.041921397379916,
+ "grad_norm": 0.008113945834338665,
+ "learning_rate": 0.00045276535851907613,
+ "loss": 3.699582099914551,
+ "step": 2739
+ },
+ {
+ "epoch": 38.05589519650655,
+ "grad_norm": 0.0072370911948382854,
+ "learning_rate": 0.0004518343227125807,
+ "loss": 3.698126792907715,
+ "step": 2740
+ },
+ {
+ "epoch": 38.069868995633186,
+ "grad_norm": 0.00817187875509262,
+ "learning_rate": 0.0004509041434794685,
+ "loss": 3.738393783569336,
+ "step": 2741
+ },
+ {
+ "epoch": 38.083842794759825,
+ "grad_norm": 0.007495569530874491,
+ "learning_rate": 0.00044997482156916845,
+ "loss": 3.727083683013916,
+ "step": 2742
+ },
+ {
+ "epoch": 38.09781659388646,
+ "grad_norm": 0.007855897769331932,
+ "learning_rate": 0.0004490463577304171,
+ "loss": 3.716137170791626,
+ "step": 2743
+ },
+ {
+ "epoch": 38.1117903930131,
+ "grad_norm": 0.007817867211997509,
+ "learning_rate": 0.0004481187527112628,
+ "loss": 3.750791072845459,
+ "step": 2744
+ },
+ {
+ "epoch": 38.12576419213974,
+ "grad_norm": 0.007712380029261112,
+ "learning_rate": 0.00044719200725905887,
+ "loss": 3.720554828643799,
+ "step": 2745
+ },
+ {
+ "epoch": 38.13973799126637,
+ "grad_norm": 0.00859921332448721,
+ "learning_rate": 0.00044626612212046816,
+ "loss": 3.742496967315674,
+ "step": 2746
+ },
+ {
+ "epoch": 38.15371179039301,
+ "grad_norm": 0.007692081853747368,
+ "learning_rate": 0.0004453410980414582,
+ "loss": 3.703803300857544,
+ "step": 2747
+ },
+ {
+ "epoch": 38.16768558951965,
+ "grad_norm": 0.007734883576631546,
+ "learning_rate": 0.00044441693576730565,
+ "loss": 3.7371301651000977,
+ "step": 2748
+ },
+ {
+ "epoch": 38.18165938864629,
+ "grad_norm": 0.009022004902362823,
+ "learning_rate": 0.0004434936360425916,
+ "loss": 3.727294683456421,
+ "step": 2749
+ },
+ {
+ "epoch": 38.19563318777293,
+ "grad_norm": 0.008058780804276466,
+ "learning_rate": 0.00044257119961120067,
+ "loss": 3.7583885192871094,
+ "step": 2750
+ },
+ {
+ "epoch": 38.209606986899566,
+ "grad_norm": 0.00698013324290514,
+ "learning_rate": 0.00044164962721632376,
+ "loss": 3.7312440872192383,
+ "step": 2751
+ },
+ {
+ "epoch": 38.223580786026204,
+ "grad_norm": 0.007999536581337452,
+ "learning_rate": 0.00044072891960045523,
+ "loss": 3.6972708702087402,
+ "step": 2752
+ },
+ {
+ "epoch": 38.237554585152836,
+ "grad_norm": 0.008125107735395432,
+ "learning_rate": 0.0004398090775053929,
+ "loss": 3.7123398780822754,
+ "step": 2753
+ },
+ {
+ "epoch": 38.251528384279474,
+ "grad_norm": 0.006736223120242357,
+ "learning_rate": 0.0004388901016722357,
+ "loss": 3.749678134918213,
+ "step": 2754
+ },
+ {
+ "epoch": 38.26550218340611,
+ "grad_norm": 0.009154478088021278,
+ "learning_rate": 0.000437971992841387,
+ "loss": 3.722728967666626,
+ "step": 2755
+ },
+ {
+ "epoch": 38.27947598253275,
+ "grad_norm": 0.008070871233940125,
+ "learning_rate": 0.00043705475175255087,
+ "loss": 3.7297017574310303,
+ "step": 2756
+ },
+ {
+ "epoch": 38.29344978165939,
+ "grad_norm": 0.007498566992580891,
+ "learning_rate": 0.0004361383791447307,
+ "loss": 3.7286131381988525,
+ "step": 2757
+ },
+ {
+ "epoch": 38.30742358078603,
+ "grad_norm": 0.007642887998372316,
+ "learning_rate": 0.00043522287575623183,
+ "loss": 3.750898838043213,
+ "step": 2758
+ },
+ {
+ "epoch": 38.32139737991266,
+ "grad_norm": 0.007800443097949028,
+ "learning_rate": 0.0004343082423246588,
+ "loss": 3.719512462615967,
+ "step": 2759
+ },
+ {
+ "epoch": 38.3353711790393,
+ "grad_norm": 0.007393732666969299,
+ "learning_rate": 0.0004333944795869159,
+ "loss": 3.7136006355285645,
+ "step": 2760
+ },
+ {
+ "epoch": 38.34934497816594,
+ "grad_norm": 0.008559555746614933,
+ "learning_rate": 0.0004324815882792042,
+ "loss": 3.7087020874023438,
+ "step": 2761
+ },
+ {
+ "epoch": 38.36331877729258,
+ "grad_norm": 0.007379850372672081,
+ "learning_rate": 0.0004315695691370241,
+ "loss": 3.7638165950775146,
+ "step": 2762
+ },
+ {
+ "epoch": 38.377292576419215,
+ "grad_norm": 0.007374506443738937,
+ "learning_rate": 0.00043065842289517303,
+ "loss": 3.7208821773529053,
+ "step": 2763
+ },
+ {
+ "epoch": 38.391266375545854,
+ "grad_norm": 0.00755207147449255,
+ "learning_rate": 0.0004297481502877451,
+ "loss": 3.738490581512451,
+ "step": 2764
+ },
+ {
+ "epoch": 38.40524017467249,
+ "grad_norm": 0.007752610836178064,
+ "learning_rate": 0.0004288387520481306,
+ "loss": 3.7401833534240723,
+ "step": 2765
+ },
+ {
+ "epoch": 38.419213973799124,
+ "grad_norm": 0.006907653529196978,
+ "learning_rate": 0.0004279302289090151,
+ "loss": 3.7754998207092285,
+ "step": 2766
+ },
+ {
+ "epoch": 38.43318777292576,
+ "grad_norm": 0.007771312724798918,
+ "learning_rate": 0.0004270225816023797,
+ "loss": 3.73626446723938,
+ "step": 2767
+ },
+ {
+ "epoch": 38.4471615720524,
+ "grad_norm": 0.007395982276648283,
+ "learning_rate": 0.0004261158108594987,
+ "loss": 3.7159175872802734,
+ "step": 2768
+ },
+ {
+ "epoch": 38.46113537117904,
+ "grad_norm": 0.007196040358394384,
+ "learning_rate": 0.0004252099174109411,
+ "loss": 3.750246047973633,
+ "step": 2769
+ },
+ {
+ "epoch": 38.47510917030568,
+ "grad_norm": 0.008241173811256886,
+ "learning_rate": 0.0004243049019865692,
+ "loss": 3.70859694480896,
+ "step": 2770
+ },
+ {
+ "epoch": 38.48908296943232,
+ "grad_norm": 0.0072947596199810505,
+ "learning_rate": 0.00042340076531553755,
+ "loss": 3.7385268211364746,
+ "step": 2771
+ },
+ {
+ "epoch": 38.50305676855895,
+ "grad_norm": 0.00782464537769556,
+ "learning_rate": 0.00042249750812629297,
+ "loss": 3.75885272026062,
+ "step": 2772
+ },
+ {
+ "epoch": 38.51703056768559,
+ "grad_norm": 0.007813232019543648,
+ "learning_rate": 0.0004215951311465722,
+ "loss": 3.71810245513916,
+ "step": 2773
+ },
+ {
+ "epoch": 38.531004366812226,
+ "grad_norm": 0.006900500506162643,
+ "learning_rate": 0.00042069363510340604,
+ "loss": 3.7365174293518066,
+ "step": 2774
+ },
+ {
+ "epoch": 38.544978165938865,
+ "grad_norm": 0.007812316995114088,
+ "learning_rate": 0.0004197930207231119,
+ "loss": 3.774794101715088,
+ "step": 2775
+ },
+ {
+ "epoch": 38.5589519650655,
+ "grad_norm": 0.007018622010946274,
+ "learning_rate": 0.00041889328873129933,
+ "loss": 3.7112040519714355,
+ "step": 2776
+ },
+ {
+ "epoch": 38.57292576419214,
+ "grad_norm": 0.006895892322063446,
+ "learning_rate": 0.0004179944398528659,
+ "loss": 3.7497243881225586,
+ "step": 2777
+ },
+ {
+ "epoch": 38.58689956331878,
+ "grad_norm": 0.0071940235793590546,
+ "learning_rate": 0.0004170964748119979,
+ "loss": 3.7566123008728027,
+ "step": 2778
+ },
+ {
+ "epoch": 38.60087336244541,
+ "grad_norm": 0.006776201073080301,
+ "learning_rate": 0.0004161993943321702,
+ "loss": 3.7683968544006348,
+ "step": 2779
+ },
+ {
+ "epoch": 38.61484716157205,
+ "grad_norm": 0.006886148825287819,
+ "learning_rate": 0.0004153031991361428,
+ "loss": 3.759763240814209,
+ "step": 2780
+ },
+ {
+ "epoch": 38.62882096069869,
+ "grad_norm": 0.006466342136263847,
+ "learning_rate": 0.0004144078899459659,
+ "loss": 3.736006498336792,
+ "step": 2781
+ },
+ {
+ "epoch": 38.64279475982533,
+ "grad_norm": 0.007055724039673805,
+ "learning_rate": 0.0004135134674829725,
+ "loss": 3.70645809173584,
+ "step": 2782
+ },
+ {
+ "epoch": 38.65676855895197,
+ "grad_norm": 0.006948428228497505,
+ "learning_rate": 0.00041261993246778303,
+ "loss": 3.7365002632141113,
+ "step": 2783
+ },
+ {
+ "epoch": 38.670742358078606,
+ "grad_norm": 0.006491438951343298,
+ "learning_rate": 0.00041172728562030255,
+ "loss": 3.735001564025879,
+ "step": 2784
+ },
+ {
+ "epoch": 38.68471615720524,
+ "grad_norm": 0.006851755082607269,
+ "learning_rate": 0.0004108355276597209,
+ "loss": 3.720893383026123,
+ "step": 2785
+ },
+ {
+ "epoch": 38.698689956331876,
+ "grad_norm": 0.007266475819051266,
+ "learning_rate": 0.00040994465930451114,
+ "loss": 3.7746100425720215,
+ "step": 2786
+ },
+ {
+ "epoch": 38.712663755458514,
+ "grad_norm": 0.007112737279385328,
+ "learning_rate": 0.00040905468127242926,
+ "loss": 3.7441720962524414,
+ "step": 2787
+ },
+ {
+ "epoch": 38.72663755458515,
+ "grad_norm": 0.006514255423098803,
+ "learning_rate": 0.000408165594280515,
+ "loss": 3.7333898544311523,
+ "step": 2788
+ },
+ {
+ "epoch": 38.74061135371179,
+ "grad_norm": 0.007172788958996534,
+ "learning_rate": 0.0004072773990450896,
+ "loss": 3.7675576210021973,
+ "step": 2789
+ },
+ {
+ "epoch": 38.75458515283843,
+ "grad_norm": 0.00684684282168746,
+ "learning_rate": 0.0004063900962817564,
+ "loss": 3.7141308784484863,
+ "step": 2790
+ },
+ {
+ "epoch": 38.76855895196506,
+ "grad_norm": 0.006578145083039999,
+ "learning_rate": 0.00040550368670539777,
+ "loss": 3.7256317138671875,
+ "step": 2791
+ },
+ {
+ "epoch": 38.7825327510917,
+ "grad_norm": 0.007047492545098066,
+ "learning_rate": 0.00040461817103017926,
+ "loss": 3.7618556022644043,
+ "step": 2792
+ },
+ {
+ "epoch": 38.79650655021834,
+ "grad_norm": 0.007036720868200064,
+ "learning_rate": 0.00040373354996954487,
+ "loss": 3.7397944927215576,
+ "step": 2793
+ },
+ {
+ "epoch": 38.81048034934498,
+ "grad_norm": 0.00745626725256443,
+ "learning_rate": 0.00040284982423621714,
+ "loss": 3.7831153869628906,
+ "step": 2794
+ },
+ {
+ "epoch": 38.82445414847162,
+ "grad_norm": 0.007454817183315754,
+ "learning_rate": 0.00040196699454219797,
+ "loss": 3.7409911155700684,
+ "step": 2795
+ },
+ {
+ "epoch": 38.838427947598255,
+ "grad_norm": 0.0075650583021342754,
+ "learning_rate": 0.00040108506159876756,
+ "loss": 3.7074005603790283,
+ "step": 2796
+ },
+ {
+ "epoch": 38.852401746724894,
+ "grad_norm": 0.006908242590725422,
+ "learning_rate": 0.0004002040261164838,
+ "loss": 3.767744779586792,
+ "step": 2797
+ },
+ {
+ "epoch": 38.866375545851525,
+ "grad_norm": 0.0072857895866036415,
+ "learning_rate": 0.0003993238888051801,
+ "loss": 3.7349488735198975,
+ "step": 2798
+ },
+ {
+ "epoch": 38.880349344978164,
+ "grad_norm": 0.0072942771948874,
+ "learning_rate": 0.00039844465037396716,
+ "loss": 3.75968861579895,
+ "step": 2799
+ },
+ {
+ "epoch": 38.8943231441048,
+ "grad_norm": 0.0071496241725981236,
+ "learning_rate": 0.00039756631153123296,
+ "loss": 3.8237380981445312,
+ "step": 2800
+ },
+ {
+ "epoch": 38.90829694323144,
+ "grad_norm": 0.006803249940276146,
+ "learning_rate": 0.00039668887298463783,
+ "loss": 3.7468650341033936,
+ "step": 2801
+ },
+ {
+ "epoch": 38.92227074235808,
+ "grad_norm": 0.007433352991938591,
+ "learning_rate": 0.00039581233544111867,
+ "loss": 3.783236503601074,
+ "step": 2802
+ },
+ {
+ "epoch": 38.93624454148472,
+ "grad_norm": 0.006633558310568333,
+ "learning_rate": 0.0003949366996068858,
+ "loss": 3.75020694732666,
+ "step": 2803
+ },
+ {
+ "epoch": 38.95021834061135,
+ "grad_norm": 0.007415154483169317,
+ "learning_rate": 0.00039406196618742376,
+ "loss": 3.7545979022979736,
+ "step": 2804
+ },
+ {
+ "epoch": 38.96419213973799,
+ "grad_norm": 0.006997843272984028,
+ "learning_rate": 0.0003931881358874886,
+ "loss": 3.7404396533966064,
+ "step": 2805
+ },
+ {
+ "epoch": 38.97816593886463,
+ "grad_norm": 0.007417330518364906,
+ "learning_rate": 0.0003923152094111095,
+ "loss": 3.7575364112854004,
+ "step": 2806
+ },
+ {
+ "epoch": 38.992139737991266,
+ "grad_norm": 0.006936701480299234,
+ "learning_rate": 0.0003914431874615889,
+ "loss": 3.771845817565918,
+ "step": 2807
+ },
+ {
+ "epoch": 39.0,
+ "grad_norm": 0.008293208666145802,
+ "learning_rate": 0.00039057207074149795,
+ "loss": 3.746669292449951,
+ "step": 2808
+ },
+ {
+ "epoch": 39.0,
+ "eval_loss": 4.370476245880127,
+ "eval_runtime": 50.9368,
+ "eval_samples_per_second": 47.942,
+ "eval_steps_per_second": 1.512,
+ "step": 2808
+ },
+ {
+ "epoch": 39.01397379912664,
+ "grad_norm": 0.0077544888481497765,
+ "learning_rate": 0.0003897018599526799,
+ "loss": 3.6915597915649414,
+ "step": 2809
+ },
+ {
+ "epoch": 39.02794759825328,
+ "grad_norm": 0.008674043230712414,
+ "learning_rate": 0.000388832555796248,
+ "loss": 3.67698073387146,
+ "step": 2810
+ },
+ {
+ "epoch": 39.041921397379916,
+ "grad_norm": 0.00724575063213706,
+ "learning_rate": 0.00038796415897258523,
+ "loss": 3.6838502883911133,
+ "step": 2811
+ },
+ {
+ "epoch": 39.05589519650655,
+ "grad_norm": 0.00849415734410286,
+ "learning_rate": 0.00038709667018134246,
+ "loss": 3.69064998626709,
+ "step": 2812
+ },
+ {
+ "epoch": 39.069868995633186,
+ "grad_norm": 0.00769044877961278,
+ "learning_rate": 0.0003862300901214403,
+ "loss": 3.670502185821533,
+ "step": 2813
+ },
+ {
+ "epoch": 39.083842794759825,
+ "grad_norm": 0.007846124470233917,
+ "learning_rate": 0.00038536441949106647,
+ "loss": 3.767771005630493,
+ "step": 2814
+ },
+ {
+ "epoch": 39.09781659388646,
+ "grad_norm": 0.007555077783763409,
+ "learning_rate": 0.000384499658987676,
+ "loss": 3.7001633644104004,
+ "step": 2815
+ },
+ {
+ "epoch": 39.1117903930131,
+ "grad_norm": 0.007834251970052719,
+ "learning_rate": 0.0003836358093079915,
+ "loss": 3.704331874847412,
+ "step": 2816
+ },
+ {
+ "epoch": 39.12576419213974,
+ "grad_norm": 0.007608097977936268,
+ "learning_rate": 0.00038277287114799974,
+ "loss": 3.7108519077301025,
+ "step": 2817
+ },
+ {
+ "epoch": 39.13973799126637,
+ "grad_norm": 0.007420446258038282,
+ "learning_rate": 0.0003819108452029561,
+ "loss": 3.697843313217163,
+ "step": 2818
+ },
+ {
+ "epoch": 39.15371179039301,
+ "grad_norm": 0.008091001771390438,
+ "learning_rate": 0.000381049732167378,
+ "loss": 3.7380661964416504,
+ "step": 2819
+ },
+ {
+ "epoch": 39.16768558951965,
+ "grad_norm": 0.007265515625476837,
+ "learning_rate": 0.00038018953273504907,
+ "loss": 3.680748462677002,
+ "step": 2820
+ },
+ {
+ "epoch": 39.18165938864629,
+ "grad_norm": 0.007735357619822025,
+ "learning_rate": 0.00037933024759901636,
+ "loss": 3.6648077964782715,
+ "step": 2821
+ },
+ {
+ "epoch": 39.19563318777293,
+ "grad_norm": 0.008188895881175995,
+ "learning_rate": 0.0003784718774515907,
+ "loss": 3.706609010696411,
+ "step": 2822
+ },
+ {
+ "epoch": 39.209606986899566,
+ "grad_norm": 0.00816027820110321,
+ "learning_rate": 0.00037761442298434554,
+ "loss": 3.7368297576904297,
+ "step": 2823
+ },
+ {
+ "epoch": 39.223580786026204,
+ "grad_norm": 0.006931936834007502,
+ "learning_rate": 0.00037675788488811566,
+ "loss": 3.7121026515960693,
+ "step": 2824
+ },
+ {
+ "epoch": 39.237554585152836,
+ "grad_norm": 0.007615855894982815,
+ "learning_rate": 0.00037590226385299897,
+ "loss": 3.7379372119903564,
+ "step": 2825
+ },
+ {
+ "epoch": 39.251528384279474,
+ "grad_norm": 0.0073505607433617115,
+ "learning_rate": 0.0003750475605683535,
+ "loss": 3.7330894470214844,
+ "step": 2826
+ },
+ {
+ "epoch": 39.26550218340611,
+ "grad_norm": 0.007151979021728039,
+ "learning_rate": 0.00037419377572279897,
+ "loss": 3.713761568069458,
+ "step": 2827
+ },
+ {
+ "epoch": 39.27947598253275,
+ "grad_norm": 0.007310053799301386,
+ "learning_rate": 0.00037334091000421326,
+ "loss": 3.705451011657715,
+ "step": 2828
+ },
+ {
+ "epoch": 39.29344978165939,
+ "grad_norm": 0.0072968751192092896,
+ "learning_rate": 0.0003724889640997357,
+ "loss": 3.7550907135009766,
+ "step": 2829
+ },
+ {
+ "epoch": 39.30742358078603,
+ "grad_norm": 0.007197601720690727,
+ "learning_rate": 0.00037163793869576404,
+ "loss": 3.7346463203430176,
+ "step": 2830
+ },
+ {
+ "epoch": 39.32139737991266,
+ "grad_norm": 0.007823911495506763,
+ "learning_rate": 0.00037078783447795307,
+ "loss": 3.7257699966430664,
+ "step": 2831
+ },
+ {
+ "epoch": 39.3353711790393,
+ "grad_norm": 0.007118315435945988,
+ "learning_rate": 0.0003699386521312171,
+ "loss": 3.7014496326446533,
+ "step": 2832
+ },
+ {
+ "epoch": 39.34934497816594,
+ "grad_norm": 0.0069230347871780396,
+ "learning_rate": 0.00036909039233972666,
+ "loss": 3.6709237098693848,
+ "step": 2833
+ },
+ {
+ "epoch": 39.36331877729258,
+ "grad_norm": 0.007836844772100449,
+ "learning_rate": 0.00036824305578691,
+ "loss": 3.695725440979004,
+ "step": 2834
+ },
+ {
+ "epoch": 39.377292576419215,
+ "grad_norm": 0.007123398594558239,
+ "learning_rate": 0.00036739664315544956,
+ "loss": 3.7278594970703125,
+ "step": 2835
+ },
+ {
+ "epoch": 39.391266375545854,
+ "grad_norm": 0.007792143151164055,
+ "learning_rate": 0.00036655115512728586,
+ "loss": 3.7291927337646484,
+ "step": 2836
+ },
+ {
+ "epoch": 39.40524017467249,
+ "grad_norm": 0.007245976477861404,
+ "learning_rate": 0.00036570659238361345,
+ "loss": 3.7183046340942383,
+ "step": 2837
+ },
+ {
+ "epoch": 39.419213973799124,
+ "grad_norm": 0.007778054103255272,
+ "learning_rate": 0.0003648629556048804,
+ "loss": 3.7485761642456055,
+ "step": 2838
+ },
+ {
+ "epoch": 39.43318777292576,
+ "grad_norm": 0.007539471611380577,
+ "learning_rate": 0.00036402024547079007,
+ "loss": 3.7273335456848145,
+ "step": 2839
+ },
+ {
+ "epoch": 39.4471615720524,
+ "grad_norm": 0.007554007228463888,
+ "learning_rate": 0.0003631784626602987,
+ "loss": 3.684359073638916,
+ "step": 2840
+ },
+ {
+ "epoch": 39.46113537117904,
+ "grad_norm": 0.007862667553126812,
+ "learning_rate": 0.00036233760785161606,
+ "loss": 3.7223072052001953,
+ "step": 2841
+ },
+ {
+ "epoch": 39.47510917030568,
+ "grad_norm": 0.00869659148156643,
+ "learning_rate": 0.0003614976817222028,
+ "loss": 3.7475168704986572,
+ "step": 2842
+ },
+ {
+ "epoch": 39.48908296943232,
+ "grad_norm": 0.00696196174249053,
+ "learning_rate": 0.00036065868494877226,
+ "loss": 3.7773914337158203,
+ "step": 2843
+ },
+ {
+ "epoch": 39.50305676855895,
+ "grad_norm": 0.008952745236456394,
+ "learning_rate": 0.00035982061820729005,
+ "loss": 3.6761155128479004,
+ "step": 2844
+ },
+ {
+ "epoch": 39.51703056768559,
+ "grad_norm": 0.008341601118445396,
+ "learning_rate": 0.00035898348217297034,
+ "loss": 3.746553421020508,
+ "step": 2845
+ },
+ {
+ "epoch": 39.531004366812226,
+ "grad_norm": 0.007268146146088839,
+ "learning_rate": 0.00035814727752027885,
+ "loss": 3.6921701431274414,
+ "step": 2846
+ },
+ {
+ "epoch": 39.544978165938865,
+ "grad_norm": 0.007401647046208382,
+ "learning_rate": 0.00035731200492293064,
+ "loss": 3.722433090209961,
+ "step": 2847
+ },
+ {
+ "epoch": 39.5589519650655,
+ "grad_norm": 0.006969883106648922,
+ "learning_rate": 0.00035647766505389006,
+ "loss": 3.7449448108673096,
+ "step": 2848
+ },
+ {
+ "epoch": 39.57292576419214,
+ "grad_norm": 0.007616426330059767,
+ "learning_rate": 0.00035564425858536895,
+ "loss": 3.725113868713379,
+ "step": 2849
+ },
+ {
+ "epoch": 39.58689956331878,
+ "grad_norm": 0.006883042864501476,
+ "learning_rate": 0.00035481178618882814,
+ "loss": 3.717106342315674,
+ "step": 2850
+ },
+ {
+ "epoch": 39.60087336244541,
+ "grad_norm": 0.007106961216777563,
+ "learning_rate": 0.00035398024853497574,
+ "loss": 3.74045467376709,
+ "step": 2851
+ },
+ {
+ "epoch": 39.61484716157205,
+ "grad_norm": 0.007402827963232994,
+ "learning_rate": 0.00035314964629376664,
+ "loss": 3.7288095951080322,
+ "step": 2852
+ },
+ {
+ "epoch": 39.62882096069869,
+ "grad_norm": 0.006770852487534285,
+ "learning_rate": 0.00035231998013440227,
+ "loss": 3.7663369178771973,
+ "step": 2853
+ },
+ {
+ "epoch": 39.64279475982533,
+ "grad_norm": 0.007622968405485153,
+ "learning_rate": 0.0003514912507253287,
+ "loss": 3.747084140777588,
+ "step": 2854
+ },
+ {
+ "epoch": 39.65676855895197,
+ "grad_norm": 0.006876434665173292,
+ "learning_rate": 0.00035066345873423975,
+ "loss": 3.7591025829315186,
+ "step": 2855
+ },
+ {
+ "epoch": 39.670742358078606,
+ "grad_norm": 0.007071215659379959,
+ "learning_rate": 0.00034983660482807116,
+ "loss": 3.7583885192871094,
+ "step": 2856
+ },
+ {
+ "epoch": 39.68471615720524,
+ "grad_norm": 0.007130554411560297,
+ "learning_rate": 0.00034901068967300465,
+ "loss": 3.6950902938842773,
+ "step": 2857
+ },
+ {
+ "epoch": 39.698689956331876,
+ "grad_norm": 0.006981188431382179,
+ "learning_rate": 0.00034818571393446524,
+ "loss": 3.7390527725219727,
+ "step": 2858
+ },
+ {
+ "epoch": 39.712663755458514,
+ "grad_norm": 0.0067793820053339005,
+ "learning_rate": 0.0003473616782771209,
+ "loss": 3.703673839569092,
+ "step": 2859
+ },
+ {
+ "epoch": 39.72663755458515,
+ "grad_norm": 0.00674026133492589,
+ "learning_rate": 0.0003465385833648827,
+ "loss": 3.6988182067871094,
+ "step": 2860
+ },
+ {
+ "epoch": 39.74061135371179,
+ "grad_norm": 0.006553220096975565,
+ "learning_rate": 0.0003457164298609021,
+ "loss": 3.7367048263549805,
+ "step": 2861
+ },
+ {
+ "epoch": 39.75458515283843,
+ "grad_norm": 0.007079749833792448,
+ "learning_rate": 0.000344895218427575,
+ "loss": 3.773226022720337,
+ "step": 2862
+ },
+ {
+ "epoch": 39.76855895196506,
+ "grad_norm": 0.006091540213674307,
+ "learning_rate": 0.00034407494972653547,
+ "loss": 3.747929096221924,
+ "step": 2863
+ },
+ {
+ "epoch": 39.7825327510917,
+ "grad_norm": 0.006783034652471542,
+ "learning_rate": 0.00034325562441865935,
+ "loss": 3.7154481410980225,
+ "step": 2864
+ },
+ {
+ "epoch": 39.79650655021834,
+ "grad_norm": 0.006683558225631714,
+ "learning_rate": 0.0003424372431640625,
+ "loss": 3.751605987548828,
+ "step": 2865
+ },
+ {
+ "epoch": 39.81048034934498,
+ "grad_norm": 0.0066695502027869225,
+ "learning_rate": 0.0003416198066220997,
+ "loss": 3.7561278343200684,
+ "step": 2866
+ },
+ {
+ "epoch": 39.82445414847162,
+ "grad_norm": 0.006601935718208551,
+ "learning_rate": 0.0003408033154513655,
+ "loss": 3.754878520965576,
+ "step": 2867
+ },
+ {
+ "epoch": 39.838427947598255,
+ "grad_norm": 0.006911450996994972,
+ "learning_rate": 0.00033998777030969085,
+ "loss": 3.69533109664917,
+ "step": 2868
+ },
+ {
+ "epoch": 39.852401746724894,
+ "grad_norm": 0.0070345234125852585,
+ "learning_rate": 0.0003391731718541465,
+ "loss": 3.7328338623046875,
+ "step": 2869
+ },
+ {
+ "epoch": 39.866375545851525,
+ "grad_norm": 0.007238153368234634,
+ "learning_rate": 0.00033835952074103964,
+ "loss": 3.7370123863220215,
+ "step": 2870
+ },
+ {
+ "epoch": 39.880349344978164,
+ "grad_norm": 0.007295276038348675,
+ "learning_rate": 0.0003375468176259142,
+ "loss": 3.7123947143554688,
+ "step": 2871
+ },
+ {
+ "epoch": 39.8943231441048,
+ "grad_norm": 0.006471444386988878,
+ "learning_rate": 0.0003367350631635506,
+ "loss": 3.73347806930542,
+ "step": 2872
+ },
+ {
+ "epoch": 39.90829694323144,
+ "grad_norm": 0.007136075291782618,
+ "learning_rate": 0.00033592425800796447,
+ "loss": 3.7337799072265625,
+ "step": 2873
+ },
+ {
+ "epoch": 39.92227074235808,
+ "grad_norm": 0.0065771411173045635,
+ "learning_rate": 0.00033511440281240725,
+ "loss": 3.733506441116333,
+ "step": 2874
+ },
+ {
+ "epoch": 39.93624454148472,
+ "grad_norm": 0.006566311698406935,
+ "learning_rate": 0.0003343054982293643,
+ "loss": 3.7325632572174072,
+ "step": 2875
+ },
+ {
+ "epoch": 39.95021834061135,
+ "grad_norm": 0.006389300804585218,
+ "learning_rate": 0.00033349754491055535,
+ "loss": 3.7027206420898438,
+ "step": 2876
+ },
+ {
+ "epoch": 39.96419213973799,
+ "grad_norm": 0.006425539031624794,
+ "learning_rate": 0.00033269054350693386,
+ "loss": 3.751466989517212,
+ "step": 2877
+ },
+ {
+ "epoch": 39.97816593886463,
+ "grad_norm": 0.006735894829034805,
+ "learning_rate": 0.0003318844946686865,
+ "loss": 3.731363296508789,
+ "step": 2878
+ },
+ {
+ "epoch": 39.992139737991266,
+ "grad_norm": 0.0061791036278009415,
+ "learning_rate": 0.00033107939904523264,
+ "loss": 3.7592201232910156,
+ "step": 2879
+ },
+ {
+ "epoch": 40.0,
+ "grad_norm": 0.007923831231892109,
+ "learning_rate": 0.0003302752572852219,
+ "loss": 3.7247159481048584,
+ "step": 2880
+ },
+ {
+ "epoch": 40.0,
+ "eval_loss": 4.371238708496094,
+ "eval_runtime": 51.312,
+ "eval_samples_per_second": 47.591,
+ "eval_steps_per_second": 1.501,
+ "step": 2880
+ },
+ {
+ "epoch": 40.01397379912664,
+ "grad_norm": 0.007161229848861694,
+ "learning_rate": 0.0003294720700365385,
+ "loss": 3.6985998153686523,
+ "step": 2881
+ },
+ {
+ "epoch": 40.02794759825328,
+ "grad_norm": 0.007182417903095484,
+ "learning_rate": 0.00032866983794629457,
+ "loss": 3.7116494178771973,
+ "step": 2882
+ },
+ {
+ "epoch": 40.041921397379916,
+ "grad_norm": 0.0072847092524170876,
+ "learning_rate": 0.00032786856166083463,
+ "loss": 3.6759095191955566,
+ "step": 2883
+ },
+ {
+ "epoch": 40.05589519650655,
+ "grad_norm": 0.006694195792078972,
+ "learning_rate": 0.0003270682418257323,
+ "loss": 3.6635122299194336,
+ "step": 2884
+ },
+ {
+ "epoch": 40.069868995633186,
+ "grad_norm": 0.0073003629222512245,
+ "learning_rate": 0.0003262688790857914,
+ "loss": 3.7130517959594727,
+ "step": 2885
+ },
+ {
+ "epoch": 40.083842794759825,
+ "grad_norm": 0.006921815685927868,
+ "learning_rate": 0.0003254704740850432,
+ "loss": 3.6923139095306396,
+ "step": 2886
+ },
+ {
+ "epoch": 40.09781659388646,
+ "grad_norm": 0.007585546933114529,
+ "learning_rate": 0.00032467302746674805,
+ "loss": 3.680659770965576,
+ "step": 2887
+ },
+ {
+ "epoch": 40.1117903930131,
+ "grad_norm": 0.0069264741614460945,
+ "learning_rate": 0.0003238765398733956,
+ "loss": 3.7042646408081055,
+ "step": 2888
+ },
+ {
+ "epoch": 40.12576419213974,
+ "grad_norm": 0.007581264711916447,
+ "learning_rate": 0.0003230810119467003,
+ "loss": 3.6890318393707275,
+ "step": 2889
+ },
+ {
+ "epoch": 40.13973799126637,
+ "grad_norm": 0.007325596176087856,
+ "learning_rate": 0.00032228644432760544,
+ "loss": 3.728785753250122,
+ "step": 2890
+ },
+ {
+ "epoch": 40.15371179039301,
+ "grad_norm": 0.006920192390680313,
+ "learning_rate": 0.0003214928376562787,
+ "loss": 3.765744686126709,
+ "step": 2891
+ },
+ {
+ "epoch": 40.16768558951965,
+ "grad_norm": 0.007247452158480883,
+ "learning_rate": 0.00032070019257211596,
+ "loss": 3.699038505554199,
+ "step": 2892
+ },
+ {
+ "epoch": 40.18165938864629,
+ "grad_norm": 0.0073222690261900425,
+ "learning_rate": 0.0003199085097137361,
+ "loss": 3.6657447814941406,
+ "step": 2893
+ },
+ {
+ "epoch": 40.19563318777293,
+ "grad_norm": 0.0074598658829927444,
+ "learning_rate": 0.000319117789718984,
+ "loss": 3.6929068565368652,
+ "step": 2894
+ },
+ {
+ "epoch": 40.209606986899566,
+ "grad_norm": 0.007144949398934841,
+ "learning_rate": 0.0003183280332249286,
+ "loss": 3.7166645526885986,
+ "step": 2895
+ },
+ {
+ "epoch": 40.223580786026204,
+ "grad_norm": 0.0068069938570261,
+ "learning_rate": 0.0003175392408678621,
+ "loss": 3.650606155395508,
+ "step": 2896
+ },
+ {
+ "epoch": 40.237554585152836,
+ "grad_norm": 0.007320682052522898,
+ "learning_rate": 0.0003167514132833012,
+ "loss": 3.69647216796875,
+ "step": 2897
+ },
+ {
+ "epoch": 40.251528384279474,
+ "grad_norm": 0.007024327293038368,
+ "learning_rate": 0.00031596455110598264,
+ "loss": 3.6764140129089355,
+ "step": 2898
+ },
+ {
+ "epoch": 40.26550218340611,
+ "grad_norm": 0.007796750869601965,
+ "learning_rate": 0.0003151786549698689,
+ "loss": 3.6841444969177246,
+ "step": 2899
+ },
+ {
+ "epoch": 40.27947598253275,
+ "grad_norm": 0.007526555098593235,
+ "learning_rate": 0.0003143937255081411,
+ "loss": 3.7181622982025146,
+ "step": 2900
+ },
+ {
+ "epoch": 40.29344978165939,
+ "grad_norm": 0.00689823878929019,
+ "learning_rate": 0.0003136097633532032,
+ "loss": 3.6805999279022217,
+ "step": 2901
+ },
+ {
+ "epoch": 40.30742358078603,
+ "grad_norm": 0.007449387572705746,
+ "learning_rate": 0.00031282676913667903,
+ "loss": 3.719672679901123,
+ "step": 2902
+ },
+ {
+ "epoch": 40.32139737991266,
+ "grad_norm": 0.0072627803310751915,
+ "learning_rate": 0.00031204474348941336,
+ "loss": 3.692244291305542,
+ "step": 2903
+ },
+ {
+ "epoch": 40.3353711790393,
+ "grad_norm": 0.0073951175436377525,
+ "learning_rate": 0.00031126368704147026,
+ "loss": 3.734830379486084,
+ "step": 2904
+ },
+ {
+ "epoch": 40.34934497816594,
+ "grad_norm": 0.007486097048968077,
+ "learning_rate": 0.0003104836004221318,
+ "loss": 3.6959404945373535,
+ "step": 2905
+ },
+ {
+ "epoch": 40.36331877729258,
+ "grad_norm": 0.007235408294945955,
+ "learning_rate": 0.0003097044842599005,
+ "loss": 3.714496612548828,
+ "step": 2906
+ },
+ {
+ "epoch": 40.377292576419215,
+ "grad_norm": 0.0074761975556612015,
+ "learning_rate": 0.00030892633918249584,
+ "loss": 3.6801066398620605,
+ "step": 2907
+ },
+ {
+ "epoch": 40.391266375545854,
+ "grad_norm": 0.007354153320193291,
+ "learning_rate": 0.0003081491658168555,
+ "loss": 3.7053956985473633,
+ "step": 2908
+ },
+ {
+ "epoch": 40.40524017467249,
+ "grad_norm": 0.0077071161940693855,
+ "learning_rate": 0.0003073729647891338,
+ "loss": 3.724217653274536,
+ "step": 2909
+ },
+ {
+ "epoch": 40.419213973799124,
+ "grad_norm": 0.007286814507097006,
+ "learning_rate": 0.00030659773672470227,
+ "loss": 3.717681646347046,
+ "step": 2910
+ },
+ {
+ "epoch": 40.43318777292576,
+ "grad_norm": 0.00741427019238472,
+ "learning_rate": 0.00030582348224814803,
+ "loss": 3.7178781032562256,
+ "step": 2911
+ },
+ {
+ "epoch": 40.4471615720524,
+ "grad_norm": 0.007106068544089794,
+ "learning_rate": 0.00030505020198327355,
+ "loss": 3.671164035797119,
+ "step": 2912
+ },
+ {
+ "epoch": 40.46113537117904,
+ "grad_norm": 0.006869092117995024,
+ "learning_rate": 0.0003042778965530969,
+ "loss": 3.7310664653778076,
+ "step": 2913
+ },
+ {
+ "epoch": 40.47510917030568,
+ "grad_norm": 0.00765156140550971,
+ "learning_rate": 0.0003035065665798508,
+ "loss": 3.7417078018188477,
+ "step": 2914
+ },
+ {
+ "epoch": 40.48908296943232,
+ "grad_norm": 0.006932055577635765,
+ "learning_rate": 0.00030273621268498183,
+ "loss": 3.6812591552734375,
+ "step": 2915
+ },
+ {
+ "epoch": 40.50305676855895,
+ "grad_norm": 0.008092030882835388,
+ "learning_rate": 0.00030196683548915055,
+ "loss": 3.69986891746521,
+ "step": 2916
+ },
+ {
+ "epoch": 40.51703056768559,
+ "grad_norm": 0.00719967158511281,
+ "learning_rate": 0.00030119843561222915,
+ "loss": 3.7247209548950195,
+ "step": 2917
+ },
+ {
+ "epoch": 40.531004366812226,
+ "grad_norm": 0.007408039178699255,
+ "learning_rate": 0.00030043101367330504,
+ "loss": 3.7106001377105713,
+ "step": 2918
+ },
+ {
+ "epoch": 40.544978165938865,
+ "grad_norm": 0.007386322598904371,
+ "learning_rate": 0.00029966457029067503,
+ "loss": 3.7073440551757812,
+ "step": 2919
+ },
+ {
+ "epoch": 40.5589519650655,
+ "grad_norm": 0.007485459093004465,
+ "learning_rate": 0.00029889910608184924,
+ "loss": 3.730858325958252,
+ "step": 2920
+ },
+ {
+ "epoch": 40.57292576419214,
+ "grad_norm": 0.007635290268808603,
+ "learning_rate": 0.00029813462166354814,
+ "loss": 3.6909053325653076,
+ "step": 2921
+ },
+ {
+ "epoch": 40.58689956331878,
+ "grad_norm": 0.0069068302400410175,
+ "learning_rate": 0.0002973711176517032,
+ "loss": 3.7286834716796875,
+ "step": 2922
+ },
+ {
+ "epoch": 40.60087336244541,
+ "grad_norm": 0.00698851840570569,
+ "learning_rate": 0.0002966085946614562,
+ "loss": 3.6587166786193848,
+ "step": 2923
+ },
+ {
+ "epoch": 40.61484716157205,
+ "grad_norm": 0.007312684785574675,
+ "learning_rate": 0.0002958470533071569,
+ "loss": 3.716031789779663,
+ "step": 2924
+ },
+ {
+ "epoch": 40.62882096069869,
+ "grad_norm": 0.006896916776895523,
+ "learning_rate": 0.00029508649420236687,
+ "loss": 3.7105724811553955,
+ "step": 2925
+ },
+ {
+ "epoch": 40.64279475982533,
+ "grad_norm": 0.007188099902123213,
+ "learning_rate": 0.00029432691795985394,
+ "loss": 3.712202548980713,
+ "step": 2926
+ },
+ {
+ "epoch": 40.65676855895197,
+ "grad_norm": 0.006414467468857765,
+ "learning_rate": 0.000293568325191595,
+ "loss": 3.71462345123291,
+ "step": 2927
+ },
+ {
+ "epoch": 40.670742358078606,
+ "grad_norm": 0.006734516471624374,
+ "learning_rate": 0.00029281071650877496,
+ "loss": 3.7277159690856934,
+ "step": 2928
+ },
+ {
+ "epoch": 40.68471615720524,
+ "grad_norm": 0.0067634680308401585,
+ "learning_rate": 0.00029205409252178524,
+ "loss": 3.6998372077941895,
+ "step": 2929
+ },
+ {
+ "epoch": 40.698689956331876,
+ "grad_norm": 0.006864939816296101,
+ "learning_rate": 0.00029129845384022446,
+ "loss": 3.7656338214874268,
+ "step": 2930
+ },
+ {
+ "epoch": 40.712663755458514,
+ "grad_norm": 0.006820365786552429,
+ "learning_rate": 0.00029054380107289626,
+ "loss": 3.7018871307373047,
+ "step": 2931
+ },
+ {
+ "epoch": 40.72663755458515,
+ "grad_norm": 0.006954800803214312,
+ "learning_rate": 0.00028979013482781114,
+ "loss": 3.7410342693328857,
+ "step": 2932
+ },
+ {
+ "epoch": 40.74061135371179,
+ "grad_norm": 0.0072957612574100494,
+ "learning_rate": 0.00028903745571218453,
+ "loss": 3.6964735984802246,
+ "step": 2933
+ },
+ {
+ "epoch": 40.75458515283843,
+ "grad_norm": 0.006805056240409613,
+ "learning_rate": 0.00028828576433243675,
+ "loss": 3.6988587379455566,
+ "step": 2934
+ },
+ {
+ "epoch": 40.76855895196506,
+ "grad_norm": 0.007712861057370901,
+ "learning_rate": 0.0002875350612941908,
+ "loss": 3.73240327835083,
+ "step": 2935
+ },
+ {
+ "epoch": 40.7825327510917,
+ "grad_norm": 0.006683751940727234,
+ "learning_rate": 0.0002867853472022755,
+ "loss": 3.6936707496643066,
+ "step": 2936
+ },
+ {
+ "epoch": 40.79650655021834,
+ "grad_norm": 0.007726171053946018,
+ "learning_rate": 0.0002860366226607223,
+ "loss": 3.7266287803649902,
+ "step": 2937
+ },
+ {
+ "epoch": 40.81048034934498,
+ "grad_norm": 0.007187390234321356,
+ "learning_rate": 0.00028528888827276396,
+ "loss": 3.7381067276000977,
+ "step": 2938
+ },
+ {
+ "epoch": 40.82445414847162,
+ "grad_norm": 0.007067828439176083,
+ "learning_rate": 0.00028454214464083706,
+ "loss": 3.71884822845459,
+ "step": 2939
+ },
+ {
+ "epoch": 40.838427947598255,
+ "grad_norm": 0.007044029422104359,
+ "learning_rate": 0.00028379639236657966,
+ "loss": 3.6823341846466064,
+ "step": 2940
+ },
+ {
+ "epoch": 40.852401746724894,
+ "grad_norm": 0.007027678657323122,
+ "learning_rate": 0.000283051632050831,
+ "loss": 3.7312417030334473,
+ "step": 2941
+ },
+ {
+ "epoch": 40.866375545851525,
+ "grad_norm": 0.0076786368153989315,
+ "learning_rate": 0.00028230786429363046,
+ "loss": 3.775012493133545,
+ "step": 2942
+ },
+ {
+ "epoch": 40.880349344978164,
+ "grad_norm": 0.0067380438558757305,
+ "learning_rate": 0.000281565089694218,
+ "loss": 3.7203941345214844,
+ "step": 2943
+ },
+ {
+ "epoch": 40.8943231441048,
+ "grad_norm": 0.007500625215470791,
+ "learning_rate": 0.00028082330885103516,
+ "loss": 3.7320876121520996,
+ "step": 2944
+ },
+ {
+ "epoch": 40.90829694323144,
+ "grad_norm": 0.006601814646273851,
+ "learning_rate": 0.00028008252236172003,
+ "loss": 3.7289505004882812,
+ "step": 2945
+ },
+ {
+ "epoch": 40.92227074235808,
+ "grad_norm": 0.00755319744348526,
+ "learning_rate": 0.00027934273082311147,
+ "loss": 3.7076311111450195,
+ "step": 2946
+ },
+ {
+ "epoch": 40.93624454148472,
+ "grad_norm": 0.0068661835975945,
+ "learning_rate": 0.0002786039348312459,
+ "loss": 3.6950671672821045,
+ "step": 2947
+ },
+ {
+ "epoch": 40.95021834061135,
+ "grad_norm": 0.007409265730530024,
+ "learning_rate": 0.00027786613498135843,
+ "loss": 3.7798776626586914,
+ "step": 2948
+ },
+ {
+ "epoch": 40.96419213973799,
+ "grad_norm": 0.006714935880154371,
+ "learning_rate": 0.0002771293318678804,
+ "loss": 3.690377712249756,
+ "step": 2949
+ },
+ {
+ "epoch": 40.97816593886463,
+ "grad_norm": 0.007498119492083788,
+ "learning_rate": 0.00027639352608444087,
+ "loss": 3.741222381591797,
+ "step": 2950
+ },
+ {
+ "epoch": 40.992139737991266,
+ "grad_norm": 0.006700210273265839,
+ "learning_rate": 0.0002756587182238656,
+ "loss": 3.7065296173095703,
+ "step": 2951
+ },
+ {
+ "epoch": 41.0,
+ "grad_norm": 0.008518884889781475,
+ "learning_rate": 0.00027492490887817597,
+ "loss": 3.7010321617126465,
+ "step": 2952
+ },
+ {
+ "epoch": 41.0,
+ "eval_loss": 4.372326374053955,
+ "eval_runtime": 48.97,
+ "eval_samples_per_second": 49.867,
+ "eval_steps_per_second": 1.572,
+ "step": 2952
+ },
+ {
+ "epoch": 41.01397379912664,
+ "grad_norm": 0.007463657297194004,
+ "learning_rate": 0.00027419209863858865,
+ "loss": 3.6893815994262695,
+ "step": 2953
+ },
+ {
+ "epoch": 41.02794759825328,
+ "grad_norm": 0.0069931019097566605,
+ "learning_rate": 0.00027346028809551576,
+ "loss": 3.6620845794677734,
+ "step": 2954
+ },
+ {
+ "epoch": 41.041921397379916,
+ "grad_norm": 0.0076762037351727486,
+ "learning_rate": 0.00027272947783856433,
+ "loss": 3.715946674346924,
+ "step": 2955
+ },
+ {
+ "epoch": 41.05589519650655,
+ "grad_norm": 0.007143040653318167,
+ "learning_rate": 0.0002719996684565341,
+ "loss": 3.67921781539917,
+ "step": 2956
+ },
+ {
+ "epoch": 41.069868995633186,
+ "grad_norm": 0.007577868644148111,
+ "learning_rate": 0.00027127086053741964,
+ "loss": 3.6886801719665527,
+ "step": 2957
+ },
+ {
+ "epoch": 41.083842794759825,
+ "grad_norm": 0.007018061354756355,
+ "learning_rate": 0.0002705430546684085,
+ "loss": 3.673877716064453,
+ "step": 2958
+ },
+ {
+ "epoch": 41.09781659388646,
+ "grad_norm": 0.0072300247848033905,
+ "learning_rate": 0.0002698162514358805,
+ "loss": 3.730193614959717,
+ "step": 2959
+ },
+ {
+ "epoch": 41.1117903930131,
+ "grad_norm": 0.00714140385389328,
+ "learning_rate": 0.00026909045142540843,
+ "loss": 3.6612749099731445,
+ "step": 2960
+ },
+ {
+ "epoch": 41.12576419213974,
+ "grad_norm": 0.0071529485285282135,
+ "learning_rate": 0.0002683656552217553,
+ "loss": 3.641969680786133,
+ "step": 2961
+ },
+ {
+ "epoch": 41.13973799126637,
+ "grad_norm": 0.00726350536569953,
+ "learning_rate": 0.0002676418634088772,
+ "loss": 3.697068452835083,
+ "step": 2962
+ },
+ {
+ "epoch": 41.15371179039301,
+ "grad_norm": 0.007082062773406506,
+ "learning_rate": 0.00026691907656991963,
+ "loss": 3.6732699871063232,
+ "step": 2963
+ },
+ {
+ "epoch": 41.16768558951965,
+ "grad_norm": 0.0069284942001104355,
+ "learning_rate": 0.00026619729528721897,
+ "loss": 3.6701955795288086,
+ "step": 2964
+ },
+ {
+ "epoch": 41.18165938864629,
+ "grad_norm": 0.007283426355570555,
+ "learning_rate": 0.00026547652014230147,
+ "loss": 3.6422624588012695,
+ "step": 2965
+ },
+ {
+ "epoch": 41.19563318777293,
+ "grad_norm": 0.006717352196574211,
+ "learning_rate": 0.0002647567517158825,
+ "loss": 3.6600942611694336,
+ "step": 2966
+ },
+ {
+ "epoch": 41.209606986899566,
+ "grad_norm": 0.006842337548732758,
+ "learning_rate": 0.00026403799058786695,
+ "loss": 3.690095901489258,
+ "step": 2967
+ },
+ {
+ "epoch": 41.223580786026204,
+ "grad_norm": 0.006624994333833456,
+ "learning_rate": 0.000263320237337347,
+ "loss": 3.7145729064941406,
+ "step": 2968
+ },
+ {
+ "epoch": 41.237554585152836,
+ "grad_norm": 0.006822563707828522,
+ "learning_rate": 0.0002626034925426036,
+ "loss": 3.734589099884033,
+ "step": 2969
+ },
+ {
+ "epoch": 41.251528384279474,
+ "grad_norm": 0.006343246903270483,
+ "learning_rate": 0.0002618877567811054,
+ "loss": 3.676018238067627,
+ "step": 2970
+ },
+ {
+ "epoch": 41.26550218340611,
+ "grad_norm": 0.007226140238344669,
+ "learning_rate": 0.00026117303062950795,
+ "loss": 3.705630302429199,
+ "step": 2971
+ },
+ {
+ "epoch": 41.27947598253275,
+ "grad_norm": 0.0066663846373558044,
+ "learning_rate": 0.0002604593146636522,
+ "loss": 3.707524299621582,
+ "step": 2972
+ },
+ {
+ "epoch": 41.29344978165939,
+ "grad_norm": 0.007017293944954872,
+ "learning_rate": 0.00025974660945856726,
+ "loss": 3.654975652694702,
+ "step": 2973
+ },
+ {
+ "epoch": 41.30742358078603,
+ "grad_norm": 0.006494272034615278,
+ "learning_rate": 0.0002590349155884669,
+ "loss": 3.7084226608276367,
+ "step": 2974
+ },
+ {
+ "epoch": 41.32139737991266,
+ "grad_norm": 0.007430107332766056,
+ "learning_rate": 0.0002583242336267492,
+ "loss": 3.681856393814087,
+ "step": 2975
+ },
+ {
+ "epoch": 41.3353711790393,
+ "grad_norm": 0.0065765841864049435,
+ "learning_rate": 0.00025761456414599807,
+ "loss": 3.7148854732513428,
+ "step": 2976
+ },
+ {
+ "epoch": 41.34934497816594,
+ "grad_norm": 0.00703916372731328,
+ "learning_rate": 0.00025690590771798165,
+ "loss": 3.707977294921875,
+ "step": 2977
+ },
+ {
+ "epoch": 41.36331877729258,
+ "grad_norm": 0.006881422363221645,
+ "learning_rate": 0.0002561982649136517,
+ "loss": 3.6452829837799072,
+ "step": 2978
+ },
+ {
+ "epoch": 41.377292576419215,
+ "grad_norm": 0.006829532328993082,
+ "learning_rate": 0.00025549163630314233,
+ "loss": 3.6812632083892822,
+ "step": 2979
+ },
+ {
+ "epoch": 41.391266375545854,
+ "grad_norm": 0.0069242967292666435,
+ "learning_rate": 0.00025478602245577255,
+ "loss": 3.6699113845825195,
+ "step": 2980
+ },
+ {
+ "epoch": 41.40524017467249,
+ "grad_norm": 0.00673352275043726,
+ "learning_rate": 0.00025408142394004265,
+ "loss": 3.67539644241333,
+ "step": 2981
+ },
+ {
+ "epoch": 41.419213973799124,
+ "grad_norm": 0.006906552705913782,
+ "learning_rate": 0.0002533778413236342,
+ "loss": 3.695683479309082,
+ "step": 2982
+ },
+ {
+ "epoch": 41.43318777292576,
+ "grad_norm": 0.007096703629940748,
+ "learning_rate": 0.0002526752751734114,
+ "loss": 3.720625877380371,
+ "step": 2983
+ },
+ {
+ "epoch": 41.4471615720524,
+ "grad_norm": 0.006843909155577421,
+ "learning_rate": 0.000251973726055419,
+ "loss": 3.6780426502227783,
+ "step": 2984
+ },
+ {
+ "epoch": 41.46113537117904,
+ "grad_norm": 0.006776071153581142,
+ "learning_rate": 0.00025127319453488277,
+ "loss": 3.702991485595703,
+ "step": 2985
+ },
+ {
+ "epoch": 41.47510917030568,
+ "grad_norm": 0.006684580352157354,
+ "learning_rate": 0.00025057368117620797,
+ "loss": 3.7161166667938232,
+ "step": 2986
+ },
+ {
+ "epoch": 41.48908296943232,
+ "grad_norm": 0.006576182786375284,
+ "learning_rate": 0.00024987518654297935,
+ "loss": 3.711947202682495,
+ "step": 2987
+ },
+ {
+ "epoch": 41.50305676855895,
+ "grad_norm": 0.006553746294230223,
+ "learning_rate": 0.0002491777111979629,
+ "loss": 3.699580669403076,
+ "step": 2988
+ },
+ {
+ "epoch": 41.51703056768559,
+ "grad_norm": 0.0063622179441154,
+ "learning_rate": 0.00024848125570310067,
+ "loss": 3.660463809967041,
+ "step": 2989
+ },
+ {
+ "epoch": 41.531004366812226,
+ "grad_norm": 0.006369025446474552,
+ "learning_rate": 0.00024778582061951466,
+ "loss": 3.712718963623047,
+ "step": 2990
+ },
+ {
+ "epoch": 41.544978165938865,
+ "grad_norm": 0.0066086845472455025,
+ "learning_rate": 0.00024709140650750426,
+ "loss": 3.7337706089019775,
+ "step": 2991
+ },
+ {
+ "epoch": 41.5589519650655,
+ "grad_norm": 0.006698573008179665,
+ "learning_rate": 0.00024639801392654687,
+ "loss": 3.693399429321289,
+ "step": 2992
+ },
+ {
+ "epoch": 41.57292576419214,
+ "grad_norm": 0.006572197191417217,
+ "learning_rate": 0.0002457056434352956,
+ "loss": 3.711644411087036,
+ "step": 2993
+ },
+ {
+ "epoch": 41.58689956331878,
+ "grad_norm": 0.0066736009903252125,
+ "learning_rate": 0.0002450142955915812,
+ "loss": 3.706545352935791,
+ "step": 2994
+ },
+ {
+ "epoch": 41.60087336244541,
+ "grad_norm": 0.006779580842703581,
+ "learning_rate": 0.0002443239709524101,
+ "loss": 3.6896963119506836,
+ "step": 2995
+ },
+ {
+ "epoch": 41.61484716157205,
+ "grad_norm": 0.0064375437796115875,
+ "learning_rate": 0.0002436346700739646,
+ "loss": 3.646042823791504,
+ "step": 2996
+ },
+ {
+ "epoch": 41.62882096069869,
+ "grad_norm": 0.006759198382496834,
+ "learning_rate": 0.00024294639351160198,
+ "loss": 3.6894006729125977,
+ "step": 2997
+ },
+ {
+ "epoch": 41.64279475982533,
+ "grad_norm": 0.00644648214802146,
+ "learning_rate": 0.00024225914181985343,
+ "loss": 3.7088422775268555,
+ "step": 2998
+ },
+ {
+ "epoch": 41.65676855895197,
+ "grad_norm": 0.00702979601919651,
+ "learning_rate": 0.00024157291555242663,
+ "loss": 3.705204486846924,
+ "step": 2999
+ },
+ {
+ "epoch": 41.670742358078606,
+ "grad_norm": 0.006393326912075281,
+ "learning_rate": 0.00024088771526220045,
+ "loss": 3.738778829574585,
+ "step": 3000
+ },
+ {
+ "epoch": 41.68471615720524,
+ "grad_norm": 0.006522264331579208,
+ "learning_rate": 0.00024020354150122884,
+ "loss": 3.7124788761138916,
+ "step": 3001
+ },
+ {
+ "epoch": 41.698689956331876,
+ "grad_norm": 0.006918660830706358,
+ "learning_rate": 0.0002395203948207383,
+ "loss": 3.703273296356201,
+ "step": 3002
+ },
+ {
+ "epoch": 41.712663755458514,
+ "grad_norm": 0.0067405398003757,
+ "learning_rate": 0.00023883827577112786,
+ "loss": 3.735112190246582,
+ "step": 3003
+ },
+ {
+ "epoch": 41.72663755458515,
+ "grad_norm": 0.006900794804096222,
+ "learning_rate": 0.00023815718490196856,
+ "loss": 3.707573413848877,
+ "step": 3004
+ },
+ {
+ "epoch": 41.74061135371179,
+ "grad_norm": 0.006647802423685789,
+ "learning_rate": 0.0002374771227620022,
+ "loss": 3.7194061279296875,
+ "step": 3005
+ },
+ {
+ "epoch": 41.75458515283843,
+ "grad_norm": 0.006786358542740345,
+ "learning_rate": 0.00023679808989914383,
+ "loss": 3.684828758239746,
+ "step": 3006
+ },
+ {
+ "epoch": 41.76855895196506,
+ "grad_norm": 0.006440348923206329,
+ "learning_rate": 0.0002361200868604771,
+ "loss": 3.679628610610962,
+ "step": 3007
+ },
+ {
+ "epoch": 41.7825327510917,
+ "grad_norm": 0.006847103592008352,
+ "learning_rate": 0.0002354431141922572,
+ "loss": 3.7070424556732178,
+ "step": 3008
+ },
+ {
+ "epoch": 41.79650655021834,
+ "grad_norm": 0.00643899803981185,
+ "learning_rate": 0.00023476717243990815,
+ "loss": 3.7474443912506104,
+ "step": 3009
+ },
+ {
+ "epoch": 41.81048034934498,
+ "grad_norm": 0.0070899054408073425,
+ "learning_rate": 0.00023409226214802462,
+ "loss": 3.6964569091796875,
+ "step": 3010
+ },
+ {
+ "epoch": 41.82445414847162,
+ "grad_norm": 0.006639422383159399,
+ "learning_rate": 0.00023341838386036994,
+ "loss": 3.7240355014801025,
+ "step": 3011
+ },
+ {
+ "epoch": 41.838427947598255,
+ "grad_norm": 0.0067732552997767925,
+ "learning_rate": 0.000232745538119875,
+ "loss": 3.691311836242676,
+ "step": 3012
+ },
+ {
+ "epoch": 41.852401746724894,
+ "grad_norm": 0.006991941947489977,
+ "learning_rate": 0.0002320737254686396,
+ "loss": 3.703491449356079,
+ "step": 3013
+ },
+ {
+ "epoch": 41.866375545851525,
+ "grad_norm": 0.006394840776920319,
+ "learning_rate": 0.0002314029464479313,
+ "loss": 3.6963045597076416,
+ "step": 3014
+ },
+ {
+ "epoch": 41.880349344978164,
+ "grad_norm": 0.00672922981902957,
+ "learning_rate": 0.00023073320159818463,
+ "loss": 3.6913256645202637,
+ "step": 3015
+ },
+ {
+ "epoch": 41.8943231441048,
+ "grad_norm": 0.006367729511111975,
+ "learning_rate": 0.00023006449145900022,
+ "loss": 3.712766170501709,
+ "step": 3016
+ },
+ {
+ "epoch": 41.90829694323144,
+ "grad_norm": 0.006370930001139641,
+ "learning_rate": 0.0002293968165691467,
+ "loss": 3.683473587036133,
+ "step": 3017
+ },
+ {
+ "epoch": 41.92227074235808,
+ "grad_norm": 0.007074430584907532,
+ "learning_rate": 0.00022873017746655758,
+ "loss": 3.7364859580993652,
+ "step": 3018
+ },
+ {
+ "epoch": 41.93624454148472,
+ "grad_norm": 0.00689726835116744,
+ "learning_rate": 0.00022806457468833126,
+ "loss": 3.7360360622406006,
+ "step": 3019
+ },
+ {
+ "epoch": 41.95021834061135,
+ "grad_norm": 0.006516663823276758,
+ "learning_rate": 0.0002274000087707323,
+ "loss": 3.7183632850646973,
+ "step": 3020
+ },
+ {
+ "epoch": 41.96419213973799,
+ "grad_norm": 0.006655904930084944,
+ "learning_rate": 0.00022673648024918938,
+ "loss": 3.7091434001922607,
+ "step": 3021
+ },
+ {
+ "epoch": 41.97816593886463,
+ "grad_norm": 0.006546393036842346,
+ "learning_rate": 0.00022607398965829604,
+ "loss": 3.682645320892334,
+ "step": 3022
+ },
+ {
+ "epoch": 41.992139737991266,
+ "grad_norm": 0.006579853594303131,
+ "learning_rate": 0.0002254125375318079,
+ "loss": 3.707092761993408,
+ "step": 3023
+ },
+ {
+ "epoch": 42.0,
+ "grad_norm": 0.007749224081635475,
+ "learning_rate": 0.00022475212440264524,
+ "loss": 3.6835598945617676,
+ "step": 3024
+ },
+ {
+ "epoch": 42.0,
+ "eval_loss": 4.3733673095703125,
+ "eval_runtime": 54.197,
+ "eval_samples_per_second": 45.058,
+ "eval_steps_per_second": 1.421,
+ "step": 3024
+ },
+ {
+ "epoch": 42.01397379912664,
+ "grad_norm": 0.006405573338270187,
+ "learning_rate": 0.000224092750802892,
+ "loss": 3.7344212532043457,
+ "step": 3025
+ },
+ {
+ "epoch": 42.02794759825328,
+ "grad_norm": 0.006732994224876165,
+ "learning_rate": 0.0002234344172637924,
+ "loss": 3.684095621109009,
+ "step": 3026
+ },
+ {
+ "epoch": 42.041921397379916,
+ "grad_norm": 0.00679958239197731,
+ "learning_rate": 0.00022277712431575434,
+ "loss": 3.697983741760254,
+ "step": 3027
+ },
+ {
+ "epoch": 42.05589519650655,
+ "grad_norm": 0.00670752814039588,
+ "learning_rate": 0.00022212087248834695,
+ "loss": 3.6364173889160156,
+ "step": 3028
+ },
+ {
+ "epoch": 42.069868995633186,
+ "grad_norm": 0.006356659810990095,
+ "learning_rate": 0.00022146566231030071,
+ "loss": 3.642988681793213,
+ "step": 3029
+ },
+ {
+ "epoch": 42.083842794759825,
+ "grad_norm": 0.006624842528253794,
+ "learning_rate": 0.000220811494309506,
+ "loss": 3.6682677268981934,
+ "step": 3030
+ },
+ {
+ "epoch": 42.09781659388646,
+ "grad_norm": 0.006486820988357067,
+ "learning_rate": 0.00022015836901301408,
+ "loss": 3.6703057289123535,
+ "step": 3031
+ },
+ {
+ "epoch": 42.1117903930131,
+ "grad_norm": 0.006477998569607735,
+ "learning_rate": 0.0002195062869470374,
+ "loss": 3.6891942024230957,
+ "step": 3032
+ },
+ {
+ "epoch": 42.12576419213974,
+ "grad_norm": 0.006716342642903328,
+ "learning_rate": 0.0002188552486369454,
+ "loss": 3.6837644577026367,
+ "step": 3033
+ },
+ {
+ "epoch": 42.13973799126637,
+ "grad_norm": 0.006492446642369032,
+ "learning_rate": 0.00021820525460726888,
+ "loss": 3.66949462890625,
+ "step": 3034
+ },
+ {
+ "epoch": 42.15371179039301,
+ "grad_norm": 0.0064978403970599174,
+ "learning_rate": 0.00021755630538169502,
+ "loss": 3.656477451324463,
+ "step": 3035
+ },
+ {
+ "epoch": 42.16768558951965,
+ "grad_norm": 0.006488895509392023,
+ "learning_rate": 0.000216908401483072,
+ "loss": 3.683609962463379,
+ "step": 3036
+ },
+ {
+ "epoch": 42.18165938864629,
+ "grad_norm": 0.0063864924013614655,
+ "learning_rate": 0.00021626154343340316,
+ "loss": 3.6635069847106934,
+ "step": 3037
+ },
+ {
+ "epoch": 42.19563318777293,
+ "grad_norm": 0.006423276849091053,
+ "learning_rate": 0.00021561573175385077,
+ "loss": 3.675956964492798,
+ "step": 3038
+ },
+ {
+ "epoch": 42.209606986899566,
+ "grad_norm": 0.006375681143254042,
+ "learning_rate": 0.00021497096696473364,
+ "loss": 3.662991523742676,
+ "step": 3039
+ },
+ {
+ "epoch": 42.223580786026204,
+ "grad_norm": 0.006157599855214357,
+ "learning_rate": 0.00021432724958552707,
+ "loss": 3.685154438018799,
+ "step": 3040
+ },
+ {
+ "epoch": 42.237554585152836,
+ "grad_norm": 0.0064014773815870285,
+ "learning_rate": 0.0002136845801348629,
+ "loss": 3.6482927799224854,
+ "step": 3041
+ },
+ {
+ "epoch": 42.251528384279474,
+ "grad_norm": 0.006111831869930029,
+ "learning_rate": 0.0002130429591305271,
+ "loss": 3.6458590030670166,
+ "step": 3042
+ },
+ {
+ "epoch": 42.26550218340611,
+ "grad_norm": 0.0062011126428842545,
+ "learning_rate": 0.0002124023870894639,
+ "loss": 3.6651253700256348,
+ "step": 3043
+ },
+ {
+ "epoch": 42.27947598253275,
+ "grad_norm": 0.006603785324841738,
+ "learning_rate": 0.00021176286452776909,
+ "loss": 3.6931943893432617,
+ "step": 3044
+ },
+ {
+ "epoch": 42.29344978165939,
+ "grad_norm": 0.006363562773913145,
+ "learning_rate": 0.00021112439196069516,
+ "loss": 3.7063207626342773,
+ "step": 3045
+ },
+ {
+ "epoch": 42.30742358078603,
+ "grad_norm": 0.006258734501898289,
+ "learning_rate": 0.0002104869699026478,
+ "loss": 3.689711093902588,
+ "step": 3046
+ },
+ {
+ "epoch": 42.32139737991266,
+ "grad_norm": 0.006544198840856552,
+ "learning_rate": 0.00020985059886718659,
+ "loss": 3.6803956031799316,
+ "step": 3047
+ },
+ {
+ "epoch": 42.3353711790393,
+ "grad_norm": 0.006236413959413767,
+ "learning_rate": 0.00020921527936702418,
+ "loss": 3.681662082672119,
+ "step": 3048
+ },
+ {
+ "epoch": 42.34934497816594,
+ "grad_norm": 0.006620924919843674,
+ "learning_rate": 0.00020858101191402557,
+ "loss": 3.6708180904388428,
+ "step": 3049
+ },
+ {
+ "epoch": 42.36331877729258,
+ "grad_norm": 0.006312021054327488,
+ "learning_rate": 0.00020794779701920895,
+ "loss": 3.68668270111084,
+ "step": 3050
+ },
+ {
+ "epoch": 42.377292576419215,
+ "grad_norm": 0.006595146376639605,
+ "learning_rate": 0.00020731563519274407,
+ "loss": 3.709224224090576,
+ "step": 3051
+ },
+ {
+ "epoch": 42.391266375545854,
+ "grad_norm": 0.006226524710655212,
+ "learning_rate": 0.00020668452694395234,
+ "loss": 3.688267707824707,
+ "step": 3052
+ },
+ {
+ "epoch": 42.40524017467249,
+ "grad_norm": 0.006482335273176432,
+ "learning_rate": 0.00020605447278130616,
+ "loss": 3.6906940937042236,
+ "step": 3053
+ },
+ {
+ "epoch": 42.419213973799124,
+ "grad_norm": 0.006389188580214977,
+ "learning_rate": 0.00020542547321242892,
+ "loss": 3.6813313961029053,
+ "step": 3054
+ },
+ {
+ "epoch": 42.43318777292576,
+ "grad_norm": 0.006554862484335899,
+ "learning_rate": 0.00020479752874409457,
+ "loss": 3.645631790161133,
+ "step": 3055
+ },
+ {
+ "epoch": 42.4471615720524,
+ "grad_norm": 0.00646898802369833,
+ "learning_rate": 0.0002041706398822257,
+ "loss": 3.691580295562744,
+ "step": 3056
+ },
+ {
+ "epoch": 42.46113537117904,
+ "grad_norm": 0.006872115656733513,
+ "learning_rate": 0.0002035448071318961,
+ "loss": 3.716787815093994,
+ "step": 3057
+ },
+ {
+ "epoch": 42.47510917030568,
+ "grad_norm": 0.006251885090023279,
+ "learning_rate": 0.00020292003099732744,
+ "loss": 3.666078567504883,
+ "step": 3058
+ },
+ {
+ "epoch": 42.48908296943232,
+ "grad_norm": 0.006771737243980169,
+ "learning_rate": 0.00020229631198189084,
+ "loss": 3.621438503265381,
+ "step": 3059
+ },
+ {
+ "epoch": 42.50305676855895,
+ "grad_norm": 0.00606488948687911,
+ "learning_rate": 0.0002016736505881058,
+ "loss": 3.6710901260375977,
+ "step": 3060
+ },
+ {
+ "epoch": 42.51703056768559,
+ "grad_norm": 0.006203590892255306,
+ "learning_rate": 0.00020105204731763832,
+ "loss": 3.6667685508728027,
+ "step": 3061
+ },
+ {
+ "epoch": 42.531004366812226,
+ "grad_norm": 0.006625751499086618,
+ "learning_rate": 0.00020043150267130412,
+ "loss": 3.648806095123291,
+ "step": 3062
+ },
+ {
+ "epoch": 42.544978165938865,
+ "grad_norm": 0.006358640734106302,
+ "learning_rate": 0.000199812017149064,
+ "loss": 3.7056827545166016,
+ "step": 3063
+ },
+ {
+ "epoch": 42.5589519650655,
+ "grad_norm": 0.006539507769048214,
+ "learning_rate": 0.00019919359125002647,
+ "loss": 3.6906323432922363,
+ "step": 3064
+ },
+ {
+ "epoch": 42.57292576419214,
+ "grad_norm": 0.007050348911434412,
+ "learning_rate": 0.00019857622547244627,
+ "loss": 3.6898245811462402,
+ "step": 3065
+ },
+ {
+ "epoch": 42.58689956331878,
+ "grad_norm": 0.006209646817296743,
+ "learning_rate": 0.00019795992031372358,
+ "loss": 3.6794214248657227,
+ "step": 3066
+ },
+ {
+ "epoch": 42.60087336244541,
+ "grad_norm": 0.0064559150487184525,
+ "learning_rate": 0.00019734467627040487,
+ "loss": 3.6937336921691895,
+ "step": 3067
+ },
+ {
+ "epoch": 42.61484716157205,
+ "grad_norm": 0.006774613168090582,
+ "learning_rate": 0.00019673049383818025,
+ "loss": 3.7005720138549805,
+ "step": 3068
+ },
+ {
+ "epoch": 42.62882096069869,
+ "grad_norm": 0.0061691561713814735,
+ "learning_rate": 0.00019611737351188642,
+ "loss": 3.7170605659484863,
+ "step": 3069
+ },
+ {
+ "epoch": 42.64279475982533,
+ "grad_norm": 0.0070815603248775005,
+ "learning_rate": 0.0001955053157855027,
+ "loss": 3.722456693649292,
+ "step": 3070
+ },
+ {
+ "epoch": 42.65676855895197,
+ "grad_norm": 0.006411026697605848,
+ "learning_rate": 0.00019489432115215308,
+ "loss": 3.669548511505127,
+ "step": 3071
+ },
+ {
+ "epoch": 42.670742358078606,
+ "grad_norm": 0.006817718036472797,
+ "learning_rate": 0.00019428439010410492,
+ "loss": 3.705488681793213,
+ "step": 3072
+ },
+ {
+ "epoch": 42.68471615720524,
+ "grad_norm": 0.006276566535234451,
+ "learning_rate": 0.0001936755231327688,
+ "loss": 3.7110562324523926,
+ "step": 3073
+ },
+ {
+ "epoch": 42.698689956331876,
+ "grad_norm": 0.006713696289807558,
+ "learning_rate": 0.0001930677207286977,
+ "loss": 3.7030649185180664,
+ "step": 3074
+ },
+ {
+ "epoch": 42.712663755458514,
+ "grad_norm": 0.006331400014460087,
+ "learning_rate": 0.00019246098338158676,
+ "loss": 3.6906652450561523,
+ "step": 3075
+ },
+ {
+ "epoch": 42.72663755458515,
+ "grad_norm": 0.007149284239858389,
+ "learning_rate": 0.00019185531158027384,
+ "loss": 3.7055468559265137,
+ "step": 3076
+ },
+ {
+ "epoch": 42.74061135371179,
+ "grad_norm": 0.006395523902028799,
+ "learning_rate": 0.00019125070581273734,
+ "loss": 3.72355318069458,
+ "step": 3077
+ },
+ {
+ "epoch": 42.75458515283843,
+ "grad_norm": 0.007234807591885328,
+ "learning_rate": 0.0001906471665660978,
+ "loss": 3.6685009002685547,
+ "step": 3078
+ },
+ {
+ "epoch": 42.76855895196506,
+ "grad_norm": 0.006079623010009527,
+ "learning_rate": 0.00019004469432661478,
+ "loss": 3.7051684856414795,
+ "step": 3079
+ },
+ {
+ "epoch": 42.7825327510917,
+ "grad_norm": 0.00699756620451808,
+ "learning_rate": 0.0001894432895796907,
+ "loss": 3.6713948249816895,
+ "step": 3080
+ },
+ {
+ "epoch": 42.79650655021834,
+ "grad_norm": 0.006799036171287298,
+ "learning_rate": 0.00018884295280986531,
+ "loss": 3.7098617553710938,
+ "step": 3081
+ },
+ {
+ "epoch": 42.81048034934498,
+ "grad_norm": 0.006591275800019503,
+ "learning_rate": 0.0001882436845008197,
+ "loss": 3.685445785522461,
+ "step": 3082
+ },
+ {
+ "epoch": 42.82445414847162,
+ "grad_norm": 0.006709378212690353,
+ "learning_rate": 0.0001876454851353733,
+ "loss": 3.6773548126220703,
+ "step": 3083
+ },
+ {
+ "epoch": 42.838427947598255,
+ "grad_norm": 0.00651182746514678,
+ "learning_rate": 0.0001870483551954847,
+ "loss": 3.6529507637023926,
+ "step": 3084
+ },
+ {
+ "epoch": 42.852401746724894,
+ "grad_norm": 0.006494257599115372,
+ "learning_rate": 0.00018645229516225093,
+ "loss": 3.692331552505493,
+ "step": 3085
+ },
+ {
+ "epoch": 42.866375545851525,
+ "grad_norm": 0.006487370003014803,
+ "learning_rate": 0.00018585730551590619,
+ "loss": 3.679516315460205,
+ "step": 3086
+ },
+ {
+ "epoch": 42.880349344978164,
+ "grad_norm": 0.006413135677576065,
+ "learning_rate": 0.00018526338673582307,
+ "loss": 3.707998752593994,
+ "step": 3087
+ },
+ {
+ "epoch": 42.8943231441048,
+ "grad_norm": 0.006509249564260244,
+ "learning_rate": 0.00018467053930051156,
+ "loss": 3.7103686332702637,
+ "step": 3088
+ },
+ {
+ "epoch": 42.90829694323144,
+ "grad_norm": 0.006331913638859987,
+ "learning_rate": 0.0001840787636876177,
+ "loss": 3.6807055473327637,
+ "step": 3089
+ },
+ {
+ "epoch": 42.92227074235808,
+ "grad_norm": 0.006528888829052448,
+ "learning_rate": 0.00018348806037392473,
+ "loss": 3.7196922302246094,
+ "step": 3090
+ },
+ {
+ "epoch": 42.93624454148472,
+ "grad_norm": 0.006212258245795965,
+ "learning_rate": 0.0001828984298353515,
+ "loss": 3.716118335723877,
+ "step": 3091
+ },
+ {
+ "epoch": 42.95021834061135,
+ "grad_norm": 0.006641864310950041,
+ "learning_rate": 0.000182309872546953,
+ "loss": 3.6726081371307373,
+ "step": 3092
+ },
+ {
+ "epoch": 42.96419213973799,
+ "grad_norm": 0.006433268543332815,
+ "learning_rate": 0.0001817223889829186,
+ "loss": 3.690727710723877,
+ "step": 3093
+ },
+ {
+ "epoch": 42.97816593886463,
+ "grad_norm": 0.00644147302955389,
+ "learning_rate": 0.00018113597961657364,
+ "loss": 3.6998791694641113,
+ "step": 3094
+ },
+ {
+ "epoch": 42.992139737991266,
+ "grad_norm": 0.006477390881627798,
+ "learning_rate": 0.00018055064492037722,
+ "loss": 3.700925350189209,
+ "step": 3095
+ },
+ {
+ "epoch": 43.0,
+ "grad_norm": 0.0077974810265004635,
+ "learning_rate": 0.00017996638536592322,
+ "loss": 3.735599994659424,
+ "step": 3096
+ },
+ {
+ "epoch": 43.0,
+ "eval_loss": 4.376750946044922,
+ "eval_runtime": 52.3576,
+ "eval_samples_per_second": 46.641,
+ "eval_steps_per_second": 1.471,
+ "step": 3096
+ },
+ {
+ "epoch": 43.01397379912664,
+ "grad_norm": 0.006609050091356039,
+ "learning_rate": 0.00017938320142393897,
+ "loss": 3.6405727863311768,
+ "step": 3097
+ },
+ {
+ "epoch": 43.02794759825328,
+ "grad_norm": 0.006470015738159418,
+ "learning_rate": 0.00017880109356428522,
+ "loss": 3.6639411449432373,
+ "step": 3098
+ },
+ {
+ "epoch": 43.041921397379916,
+ "grad_norm": 0.006597487255930901,
+ "learning_rate": 0.00017822006225595596,
+ "loss": 3.6797537803649902,
+ "step": 3099
+ },
+ {
+ "epoch": 43.05589519650655,
+ "grad_norm": 0.0064637972973287106,
+ "learning_rate": 0.0001776401079670772,
+ "loss": 3.6798789501190186,
+ "step": 3100
+ },
+ {
+ "epoch": 43.069868995633186,
+ "grad_norm": 0.006872601341456175,
+ "learning_rate": 0.00017706123116490775,
+ "loss": 3.649874210357666,
+ "step": 3101
+ },
+ {
+ "epoch": 43.083842794759825,
+ "grad_norm": 0.0064028468914330006,
+ "learning_rate": 0.00017648343231583825,
+ "loss": 3.671036720275879,
+ "step": 3102
+ },
+ {
+ "epoch": 43.09781659388646,
+ "grad_norm": 0.006735685281455517,
+ "learning_rate": 0.00017590671188539084,
+ "loss": 3.6818368434906006,
+ "step": 3103
+ },
+ {
+ "epoch": 43.1117903930131,
+ "grad_norm": 0.006652928423136473,
+ "learning_rate": 0.00017533107033821895,
+ "loss": 3.660362720489502,
+ "step": 3104
+ },
+ {
+ "epoch": 43.12576419213974,
+ "grad_norm": 0.006564146373420954,
+ "learning_rate": 0.00017475650813810598,
+ "loss": 3.6444082260131836,
+ "step": 3105
+ },
+ {
+ "epoch": 43.13973799126637,
+ "grad_norm": 0.006476754322648048,
+ "learning_rate": 0.00017418302574796716,
+ "loss": 3.6836254596710205,
+ "step": 3106
+ },
+ {
+ "epoch": 43.15371179039301,
+ "grad_norm": 0.007108788937330246,
+ "learning_rate": 0.0001736106236298462,
+ "loss": 3.6424341201782227,
+ "step": 3107
+ },
+ {
+ "epoch": 43.16768558951965,
+ "grad_norm": 0.006570066325366497,
+ "learning_rate": 0.00017303930224491735,
+ "loss": 3.663379192352295,
+ "step": 3108
+ },
+ {
+ "epoch": 43.18165938864629,
+ "grad_norm": 0.0066614942625164986,
+ "learning_rate": 0.00017246906205348395,
+ "loss": 3.6918721199035645,
+ "step": 3109
+ },
+ {
+ "epoch": 43.19563318777293,
+ "grad_norm": 0.00704583153128624,
+ "learning_rate": 0.00017189990351497824,
+ "loss": 3.692559242248535,
+ "step": 3110
+ },
+ {
+ "epoch": 43.209606986899566,
+ "grad_norm": 0.006617981940507889,
+ "learning_rate": 0.00017133182708796126,
+ "loss": 3.6591439247131348,
+ "step": 3111
+ },
+ {
+ "epoch": 43.223580786026204,
+ "grad_norm": 0.006916946265846491,
+ "learning_rate": 0.00017076483323012127,
+ "loss": 3.7165699005126953,
+ "step": 3112
+ },
+ {
+ "epoch": 43.237554585152836,
+ "grad_norm": 0.006832561455667019,
+ "learning_rate": 0.00017019892239827534,
+ "loss": 3.7023603916168213,
+ "step": 3113
+ },
+ {
+ "epoch": 43.251528384279474,
+ "grad_norm": 0.00654558464884758,
+ "learning_rate": 0.00016963409504836777,
+ "loss": 3.6702237129211426,
+ "step": 3114
+ },
+ {
+ "epoch": 43.26550218340611,
+ "grad_norm": 0.007000794634222984,
+ "learning_rate": 0.0001690703516354698,
+ "loss": 3.69088077545166,
+ "step": 3115
+ },
+ {
+ "epoch": 43.27947598253275,
+ "grad_norm": 0.006473604589700699,
+ "learning_rate": 0.00016850769261377856,
+ "loss": 3.6641716957092285,
+ "step": 3116
+ },
+ {
+ "epoch": 43.29344978165939,
+ "grad_norm": 0.006787394639104605,
+ "learning_rate": 0.000167946118436619,
+ "loss": 3.6577839851379395,
+ "step": 3117
+ },
+ {
+ "epoch": 43.30742358078603,
+ "grad_norm": 0.006220053881406784,
+ "learning_rate": 0.00016738562955644144,
+ "loss": 3.6735315322875977,
+ "step": 3118
+ },
+ {
+ "epoch": 43.32139737991266,
+ "grad_norm": 0.006513087544590235,
+ "learning_rate": 0.00016682622642482113,
+ "loss": 3.624807119369507,
+ "step": 3119
+ },
+ {
+ "epoch": 43.3353711790393,
+ "grad_norm": 0.006552594713866711,
+ "learning_rate": 0.0001662679094924592,
+ "loss": 3.679218053817749,
+ "step": 3120
+ },
+ {
+ "epoch": 43.34934497816594,
+ "grad_norm": 0.006950687617063522,
+ "learning_rate": 0.0001657106792091814,
+ "loss": 3.6864023208618164,
+ "step": 3121
+ },
+ {
+ "epoch": 43.36331877729258,
+ "grad_norm": 0.006092227529734373,
+ "learning_rate": 0.0001651545360239387,
+ "loss": 3.693777084350586,
+ "step": 3122
+ },
+ {
+ "epoch": 43.377292576419215,
+ "grad_norm": 0.006924188230186701,
+ "learning_rate": 0.0001645994803848046,
+ "loss": 3.68534779548645,
+ "step": 3123
+ },
+ {
+ "epoch": 43.391266375545854,
+ "grad_norm": 0.0067315176129341125,
+ "learning_rate": 0.00016404551273897826,
+ "loss": 3.632850170135498,
+ "step": 3124
+ },
+ {
+ "epoch": 43.40524017467249,
+ "grad_norm": 0.007006707601249218,
+ "learning_rate": 0.00016349263353278146,
+ "loss": 3.6664111614227295,
+ "step": 3125
+ },
+ {
+ "epoch": 43.419213973799124,
+ "grad_norm": 0.0067031108774244785,
+ "learning_rate": 0.00016294084321165846,
+ "loss": 3.6863315105438232,
+ "step": 3126
+ },
+ {
+ "epoch": 43.43318777292576,
+ "grad_norm": 0.006502983160316944,
+ "learning_rate": 0.00016239014222017682,
+ "loss": 3.678041458129883,
+ "step": 3127
+ },
+ {
+ "epoch": 43.4471615720524,
+ "grad_norm": 0.006409413646906614,
+ "learning_rate": 0.00016184053100202665,
+ "loss": 3.692004919052124,
+ "step": 3128
+ },
+ {
+ "epoch": 43.46113537117904,
+ "grad_norm": 0.006251887418329716,
+ "learning_rate": 0.0001612920100000198,
+ "loss": 3.64532470703125,
+ "step": 3129
+ },
+ {
+ "epoch": 43.47510917030568,
+ "grad_norm": 0.0065657407976686954,
+ "learning_rate": 0.00016074457965608942,
+ "loss": 3.638624429702759,
+ "step": 3130
+ },
+ {
+ "epoch": 43.48908296943232,
+ "grad_norm": 0.006246988195925951,
+ "learning_rate": 0.00016019824041129004,
+ "loss": 3.668254852294922,
+ "step": 3131
+ },
+ {
+ "epoch": 43.50305676855895,
+ "grad_norm": 0.0065050385892391205,
+ "learning_rate": 0.0001596529927057984,
+ "loss": 3.6901469230651855,
+ "step": 3132
+ },
+ {
+ "epoch": 43.51703056768559,
+ "grad_norm": 0.006240491755306721,
+ "learning_rate": 0.00015910883697890977,
+ "loss": 3.6796071529388428,
+ "step": 3133
+ },
+ {
+ "epoch": 43.531004366812226,
+ "grad_norm": 0.006674987263977528,
+ "learning_rate": 0.00015856577366904075,
+ "loss": 3.7001953125,
+ "step": 3134
+ },
+ {
+ "epoch": 43.544978165938865,
+ "grad_norm": 0.006260441616177559,
+ "learning_rate": 0.00015802380321372776,
+ "loss": 3.6646556854248047,
+ "step": 3135
+ },
+ {
+ "epoch": 43.5589519650655,
+ "grad_norm": 0.006696469150483608,
+ "learning_rate": 0.00015748292604962687,
+ "loss": 3.6791043281555176,
+ "step": 3136
+ },
+ {
+ "epoch": 43.57292576419214,
+ "grad_norm": 0.006716041825711727,
+ "learning_rate": 0.00015694314261251245,
+ "loss": 3.710442066192627,
+ "step": 3137
+ },
+ {
+ "epoch": 43.58689956331878,
+ "grad_norm": 0.006856406573206186,
+ "learning_rate": 0.0001564044533372785,
+ "loss": 3.6615681648254395,
+ "step": 3138
+ },
+ {
+ "epoch": 43.60087336244541,
+ "grad_norm": 0.0064330073073506355,
+ "learning_rate": 0.00015586685865793748,
+ "loss": 3.707085609436035,
+ "step": 3139
+ },
+ {
+ "epoch": 43.61484716157205,
+ "grad_norm": 0.006748909596353769,
+ "learning_rate": 0.00015533035900761955,
+ "loss": 3.6814656257629395,
+ "step": 3140
+ },
+ {
+ "epoch": 43.62882096069869,
+ "grad_norm": 0.006306119728833437,
+ "learning_rate": 0.00015479495481857282,
+ "loss": 3.6595420837402344,
+ "step": 3141
+ },
+ {
+ "epoch": 43.64279475982533,
+ "grad_norm": 0.0061928280629217625,
+ "learning_rate": 0.00015426064652216255,
+ "loss": 3.686880111694336,
+ "step": 3142
+ },
+ {
+ "epoch": 43.65676855895197,
+ "grad_norm": 0.006499494891613722,
+ "learning_rate": 0.00015372743454887173,
+ "loss": 3.6901583671569824,
+ "step": 3143
+ },
+ {
+ "epoch": 43.670742358078606,
+ "grad_norm": 0.005914116278290749,
+ "learning_rate": 0.00015319531932829917,
+ "loss": 3.699162006378174,
+ "step": 3144
+ },
+ {
+ "epoch": 43.68471615720524,
+ "grad_norm": 0.006499269045889378,
+ "learning_rate": 0.00015266430128916075,
+ "loss": 3.6740684509277344,
+ "step": 3145
+ },
+ {
+ "epoch": 43.698689956331876,
+ "grad_norm": 0.006433961447328329,
+ "learning_rate": 0.000152134380859288,
+ "loss": 3.684006929397583,
+ "step": 3146
+ },
+ {
+ "epoch": 43.712663755458514,
+ "grad_norm": 0.006057603284716606,
+ "learning_rate": 0.00015160555846562826,
+ "loss": 3.6612191200256348,
+ "step": 3147
+ },
+ {
+ "epoch": 43.72663755458515,
+ "grad_norm": 0.006337927188724279,
+ "learning_rate": 0.00015107783453424456,
+ "loss": 3.6981983184814453,
+ "step": 3148
+ },
+ {
+ "epoch": 43.74061135371179,
+ "grad_norm": 0.006348243448883295,
+ "learning_rate": 0.00015055120949031365,
+ "loss": 3.6667895317077637,
+ "step": 3149
+ },
+ {
+ "epoch": 43.75458515283843,
+ "grad_norm": 0.0060755726881325245,
+ "learning_rate": 0.00015002568375812876,
+ "loss": 3.6844797134399414,
+ "step": 3150
+ },
+ {
+ "epoch": 43.76855895196506,
+ "grad_norm": 0.006059548351913691,
+ "learning_rate": 0.0001495012577610959,
+ "loss": 3.6676461696624756,
+ "step": 3151
+ },
+ {
+ "epoch": 43.7825327510917,
+ "grad_norm": 0.006432189140468836,
+ "learning_rate": 0.00014897793192173602,
+ "loss": 3.7036523818969727,
+ "step": 3152
+ },
+ {
+ "epoch": 43.79650655021834,
+ "grad_norm": 0.006043825298547745,
+ "learning_rate": 0.00014845570666168249,
+ "loss": 3.7043371200561523,
+ "step": 3153
+ },
+ {
+ "epoch": 43.81048034934498,
+ "grad_norm": 0.0066383229568600655,
+ "learning_rate": 0.0001479345824016837,
+ "loss": 3.6830334663391113,
+ "step": 3154
+ },
+ {
+ "epoch": 43.82445414847162,
+ "grad_norm": 0.006384526379406452,
+ "learning_rate": 0.00014741455956160005,
+ "loss": 3.6746675968170166,
+ "step": 3155
+ },
+ {
+ "epoch": 43.838427947598255,
+ "grad_norm": 0.006477308925241232,
+ "learning_rate": 0.0001468956385604038,
+ "loss": 3.6788225173950195,
+ "step": 3156
+ },
+ {
+ "epoch": 43.852401746724894,
+ "grad_norm": 0.006130026653409004,
+ "learning_rate": 0.00014637781981618054,
+ "loss": 3.6821796894073486,
+ "step": 3157
+ },
+ {
+ "epoch": 43.866375545851525,
+ "grad_norm": 0.006257957778871059,
+ "learning_rate": 0.00014586110374612777,
+ "loss": 3.683966636657715,
+ "step": 3158
+ },
+ {
+ "epoch": 43.880349344978164,
+ "grad_norm": 0.006265794392675161,
+ "learning_rate": 0.00014534549076655425,
+ "loss": 3.691389560699463,
+ "step": 3159
+ },
+ {
+ "epoch": 43.8943231441048,
+ "grad_norm": 0.006294097751379013,
+ "learning_rate": 0.0001448309812928794,
+ "loss": 3.659594774246216,
+ "step": 3160
+ },
+ {
+ "epoch": 43.90829694323144,
+ "grad_norm": 0.006432943977415562,
+ "learning_rate": 0.00014431757573963504,
+ "loss": 3.665633201599121,
+ "step": 3161
+ },
+ {
+ "epoch": 43.92227074235808,
+ "grad_norm": 0.006102072540670633,
+ "learning_rate": 0.0001438052745204626,
+ "loss": 3.6348228454589844,
+ "step": 3162
+ },
+ {
+ "epoch": 43.93624454148472,
+ "grad_norm": 0.00636800192296505,
+ "learning_rate": 0.00014329407804811333,
+ "loss": 3.624040365219116,
+ "step": 3163
+ },
+ {
+ "epoch": 43.95021834061135,
+ "grad_norm": 0.006060190033167601,
+ "learning_rate": 0.00014278398673444966,
+ "loss": 3.6986045837402344,
+ "step": 3164
+ },
+ {
+ "epoch": 43.96419213973799,
+ "grad_norm": 0.0061101666651666164,
+ "learning_rate": 0.00014227500099044254,
+ "loss": 3.6816675662994385,
+ "step": 3165
+ },
+ {
+ "epoch": 43.97816593886463,
+ "grad_norm": 0.006048897281289101,
+ "learning_rate": 0.00014176712122617293,
+ "loss": 3.6739861965179443,
+ "step": 3166
+ },
+ {
+ "epoch": 43.992139737991266,
+ "grad_norm": 0.005954945459961891,
+ "learning_rate": 0.00014126034785082997,
+ "loss": 3.679354190826416,
+ "step": 3167
+ },
+ {
+ "epoch": 44.0,
+ "grad_norm": 0.007545145694166422,
+ "learning_rate": 0.0001407546812727119,
+ "loss": 3.7448782920837402,
+ "step": 3168
+ },
+ {
+ "epoch": 44.0,
+ "eval_loss": 4.377684116363525,
+ "eval_runtime": 51.0591,
+ "eval_samples_per_second": 47.827,
+ "eval_steps_per_second": 1.508,
+ "step": 3168
+ },
+ {
+ "epoch": 44.01397379912664,
+ "grad_norm": 0.00643246341496706,
+ "learning_rate": 0.00014025012189922562,
+ "loss": 3.6544783115386963,
+ "step": 3169
+ },
+ {
+ "epoch": 44.02794759825328,
+ "grad_norm": 0.006309461314231157,
+ "learning_rate": 0.00013974667013688495,
+ "loss": 3.701963186264038,
+ "step": 3170
+ },
+ {
+ "epoch": 44.041921397379916,
+ "grad_norm": 0.006177262403070927,
+ "learning_rate": 0.00013924432639131215,
+ "loss": 3.649641275405884,
+ "step": 3171
+ },
+ {
+ "epoch": 44.05589519650655,
+ "grad_norm": 0.006319771986454725,
+ "learning_rate": 0.00013874309106723636,
+ "loss": 3.67417573928833,
+ "step": 3172
+ },
+ {
+ "epoch": 44.069868995633186,
+ "grad_norm": 0.006548433564603329,
+ "learning_rate": 0.00013824296456849426,
+ "loss": 3.6257686614990234,
+ "step": 3173
+ },
+ {
+ "epoch": 44.083842794759825,
+ "grad_norm": 0.006148583721369505,
+ "learning_rate": 0.00013774394729802804,
+ "loss": 3.675339460372925,
+ "step": 3174
+ },
+ {
+ "epoch": 44.09781659388646,
+ "grad_norm": 0.0063438634388148785,
+ "learning_rate": 0.00013724603965788716,
+ "loss": 3.666316270828247,
+ "step": 3175
+ },
+ {
+ "epoch": 44.1117903930131,
+ "grad_norm": 0.006417148746550083,
+ "learning_rate": 0.00013674924204922732,
+ "loss": 3.6692237854003906,
+ "step": 3176
+ },
+ {
+ "epoch": 44.12576419213974,
+ "grad_norm": 0.0062498473562300205,
+ "learning_rate": 0.00013625355487230872,
+ "loss": 3.70363187789917,
+ "step": 3177
+ },
+ {
+ "epoch": 44.13973799126637,
+ "grad_norm": 0.006249656435102224,
+ "learning_rate": 0.00013575897852649796,
+ "loss": 3.691108226776123,
+ "step": 3178
+ },
+ {
+ "epoch": 44.15371179039301,
+ "grad_norm": 0.006240040063858032,
+ "learning_rate": 0.0001352655134102657,
+ "loss": 3.6781342029571533,
+ "step": 3179
+ },
+ {
+ "epoch": 44.16768558951965,
+ "grad_norm": 0.006186297629028559,
+ "learning_rate": 0.00013477315992118885,
+ "loss": 3.644334316253662,
+ "step": 3180
+ },
+ {
+ "epoch": 44.18165938864629,
+ "grad_norm": 0.006283515132963657,
+ "learning_rate": 0.00013428191845594675,
+ "loss": 3.6691460609436035,
+ "step": 3181
+ },
+ {
+ "epoch": 44.19563318777293,
+ "grad_norm": 0.0064705004915595055,
+ "learning_rate": 0.00013379178941032427,
+ "loss": 3.64168119430542,
+ "step": 3182
+ },
+ {
+ "epoch": 44.209606986899566,
+ "grad_norm": 0.006210929248481989,
+ "learning_rate": 0.00013330277317920959,
+ "loss": 3.66428804397583,
+ "step": 3183
+ },
+ {
+ "epoch": 44.223580786026204,
+ "grad_norm": 0.006095333956182003,
+ "learning_rate": 0.00013281487015659395,
+ "loss": 3.6205215454101562,
+ "step": 3184
+ },
+ {
+ "epoch": 44.237554585152836,
+ "grad_norm": 0.006364993751049042,
+ "learning_rate": 0.00013232808073557244,
+ "loss": 3.662804126739502,
+ "step": 3185
+ },
+ {
+ "epoch": 44.251528384279474,
+ "grad_norm": 0.006166066508740187,
+ "learning_rate": 0.00013184240530834173,
+ "loss": 3.702338218688965,
+ "step": 3186
+ },
+ {
+ "epoch": 44.26550218340611,
+ "grad_norm": 0.006147229578346014,
+ "learning_rate": 0.00013135784426620282,
+ "loss": 3.660149097442627,
+ "step": 3187
+ },
+ {
+ "epoch": 44.27947598253275,
+ "grad_norm": 0.006086066365242004,
+ "learning_rate": 0.00013087439799955674,
+ "loss": 3.652472972869873,
+ "step": 3188
+ },
+ {
+ "epoch": 44.29344978165939,
+ "grad_norm": 0.006371276918798685,
+ "learning_rate": 0.00013039206689790815,
+ "loss": 3.6471242904663086,
+ "step": 3189
+ },
+ {
+ "epoch": 44.30742358078603,
+ "grad_norm": 0.005952878389507532,
+ "learning_rate": 0.000129910851349862,
+ "loss": 3.6450729370117188,
+ "step": 3190
+ },
+ {
+ "epoch": 44.32139737991266,
+ "grad_norm": 0.006385558284819126,
+ "learning_rate": 0.00012943075174312536,
+ "loss": 3.6730473041534424,
+ "step": 3191
+ },
+ {
+ "epoch": 44.3353711790393,
+ "grad_norm": 0.005896324757486582,
+ "learning_rate": 0.00012895176846450582,
+ "loss": 3.6646409034729004,
+ "step": 3192
+ },
+ {
+ "epoch": 44.34934497816594,
+ "grad_norm": 0.006053723394870758,
+ "learning_rate": 0.00012847390189991121,
+ "loss": 3.6568682193756104,
+ "step": 3193
+ },
+ {
+ "epoch": 44.36331877729258,
+ "grad_norm": 0.005929332692176104,
+ "learning_rate": 0.00012799715243435027,
+ "loss": 3.7030997276306152,
+ "step": 3194
+ },
+ {
+ "epoch": 44.377292576419215,
+ "grad_norm": 0.006153058726340532,
+ "learning_rate": 0.00012752152045193116,
+ "loss": 3.6949849128723145,
+ "step": 3195
+ },
+ {
+ "epoch": 44.391266375545854,
+ "grad_norm": 0.006193236447870731,
+ "learning_rate": 0.0001270470063358624,
+ "loss": 3.6040501594543457,
+ "step": 3196
+ },
+ {
+ "epoch": 44.40524017467249,
+ "grad_norm": 0.005849961191415787,
+ "learning_rate": 0.0001265736104684509,
+ "loss": 3.602717399597168,
+ "step": 3197
+ },
+ {
+ "epoch": 44.419213973799124,
+ "grad_norm": 0.0060405549593269825,
+ "learning_rate": 0.00012610133323110356,
+ "loss": 3.668717861175537,
+ "step": 3198
+ },
+ {
+ "epoch": 44.43318777292576,
+ "grad_norm": 0.0062002381309866905,
+ "learning_rate": 0.00012563017500432565,
+ "loss": 3.6333389282226562,
+ "step": 3199
+ },
+ {
+ "epoch": 44.4471615720524,
+ "grad_norm": 0.006093314848840237,
+ "learning_rate": 0.00012516013616772065,
+ "loss": 3.662696361541748,
+ "step": 3200
+ },
+ {
+ "epoch": 44.46113537117904,
+ "grad_norm": 0.005765238776803017,
+ "learning_rate": 0.00012469121709999026,
+ "loss": 3.6843791007995605,
+ "step": 3201
+ },
+ {
+ "epoch": 44.47510917030568,
+ "grad_norm": 0.005989165045320988,
+ "learning_rate": 0.00012422341817893448,
+ "loss": 3.6904735565185547,
+ "step": 3202
+ },
+ {
+ "epoch": 44.48908296943232,
+ "grad_norm": 0.006150389555841684,
+ "learning_rate": 0.00012375673978145022,
+ "loss": 3.6614508628845215,
+ "step": 3203
+ },
+ {
+ "epoch": 44.50305676855895,
+ "grad_norm": 0.005898299627006054,
+ "learning_rate": 0.00012329118228353178,
+ "loss": 3.7193222045898438,
+ "step": 3204
+ },
+ {
+ "epoch": 44.51703056768559,
+ "grad_norm": 0.0058402069844305515,
+ "learning_rate": 0.00012282674606027016,
+ "loss": 3.669564723968506,
+ "step": 3205
+ },
+ {
+ "epoch": 44.531004366812226,
+ "grad_norm": 0.006015103310346603,
+ "learning_rate": 0.00012236343148585413,
+ "loss": 3.686734914779663,
+ "step": 3206
+ },
+ {
+ "epoch": 44.544978165938865,
+ "grad_norm": 0.005990280769765377,
+ "learning_rate": 0.00012190123893356703,
+ "loss": 3.664945125579834,
+ "step": 3207
+ },
+ {
+ "epoch": 44.5589519650655,
+ "grad_norm": 0.006110248621553183,
+ "learning_rate": 0.00012144016877578942,
+ "loss": 3.6546223163604736,
+ "step": 3208
+ },
+ {
+ "epoch": 44.57292576419214,
+ "grad_norm": 0.006006988696753979,
+ "learning_rate": 0.000120980221383997,
+ "loss": 3.669175386428833,
+ "step": 3209
+ },
+ {
+ "epoch": 44.58689956331878,
+ "grad_norm": 0.006008700001984835,
+ "learning_rate": 0.00012052139712876141,
+ "loss": 3.672668695449829,
+ "step": 3210
+ },
+ {
+ "epoch": 44.60087336244541,
+ "grad_norm": 0.005862198304384947,
+ "learning_rate": 0.00012006369637974875,
+ "loss": 3.6824991703033447,
+ "step": 3211
+ },
+ {
+ "epoch": 44.61484716157205,
+ "grad_norm": 0.005960221868008375,
+ "learning_rate": 0.00011960711950572005,
+ "loss": 3.6258697509765625,
+ "step": 3212
+ },
+ {
+ "epoch": 44.62882096069869,
+ "grad_norm": 0.005813480354845524,
+ "learning_rate": 0.00011915166687453174,
+ "loss": 3.6463029384613037,
+ "step": 3213
+ },
+ {
+ "epoch": 44.64279475982533,
+ "grad_norm": 0.006002446636557579,
+ "learning_rate": 0.0001186973388531332,
+ "loss": 3.721970558166504,
+ "step": 3214
+ },
+ {
+ "epoch": 44.65676855895197,
+ "grad_norm": 0.006029803771525621,
+ "learning_rate": 0.00011824413580756844,
+ "loss": 3.685335159301758,
+ "step": 3215
+ },
+ {
+ "epoch": 44.670742358078606,
+ "grad_norm": 0.005933200940489769,
+ "learning_rate": 0.0001177920581029752,
+ "loss": 3.6834492683410645,
+ "step": 3216
+ },
+ {
+ "epoch": 44.68471615720524,
+ "grad_norm": 0.005942870397120714,
+ "learning_rate": 0.00011734110610358423,
+ "loss": 3.670149326324463,
+ "step": 3217
+ },
+ {
+ "epoch": 44.698689956331876,
+ "grad_norm": 0.005972386337816715,
+ "learning_rate": 0.00011689128017271901,
+ "loss": 3.6798110008239746,
+ "step": 3218
+ },
+ {
+ "epoch": 44.712663755458514,
+ "grad_norm": 0.005964178126305342,
+ "learning_rate": 0.00011644258067279662,
+ "loss": 3.6620044708251953,
+ "step": 3219
+ },
+ {
+ "epoch": 44.72663755458515,
+ "grad_norm": 0.0059594810009002686,
+ "learning_rate": 0.00011599500796532605,
+ "loss": 3.683058738708496,
+ "step": 3220
+ },
+ {
+ "epoch": 44.74061135371179,
+ "grad_norm": 0.005901646800339222,
+ "learning_rate": 0.00011554856241090844,
+ "loss": 3.6791443824768066,
+ "step": 3221
+ },
+ {
+ "epoch": 44.75458515283843,
+ "grad_norm": 0.006178564857691526,
+ "learning_rate": 0.00011510324436923708,
+ "loss": 3.6620421409606934,
+ "step": 3222
+ },
+ {
+ "epoch": 44.76855895196506,
+ "grad_norm": 0.005973632447421551,
+ "learning_rate": 0.0001146590541990962,
+ "loss": 3.6821961402893066,
+ "step": 3223
+ },
+ {
+ "epoch": 44.7825327510917,
+ "grad_norm": 0.005914160516113043,
+ "learning_rate": 0.00011421599225836265,
+ "loss": 3.6588668823242188,
+ "step": 3224
+ },
+ {
+ "epoch": 44.79650655021834,
+ "grad_norm": 0.005887249484658241,
+ "learning_rate": 0.00011377405890400266,
+ "loss": 3.683187484741211,
+ "step": 3225
+ },
+ {
+ "epoch": 44.81048034934498,
+ "grad_norm": 0.006068149581551552,
+ "learning_rate": 0.00011333325449207415,
+ "loss": 3.6633477210998535,
+ "step": 3226
+ },
+ {
+ "epoch": 44.82445414847162,
+ "grad_norm": 0.005748802796006203,
+ "learning_rate": 0.00011289357937772547,
+ "loss": 3.676607608795166,
+ "step": 3227
+ },
+ {
+ "epoch": 44.838427947598255,
+ "grad_norm": 0.006032158620655537,
+ "learning_rate": 0.00011245503391519489,
+ "loss": 3.652369499206543,
+ "step": 3228
+ },
+ {
+ "epoch": 44.852401746724894,
+ "grad_norm": 0.00615714630112052,
+ "learning_rate": 0.00011201761845781049,
+ "loss": 3.7210354804992676,
+ "step": 3229
+ },
+ {
+ "epoch": 44.866375545851525,
+ "grad_norm": 0.0059432960115373135,
+ "learning_rate": 0.00011158133335799003,
+ "loss": 3.640336036682129,
+ "step": 3230
+ },
+ {
+ "epoch": 44.880349344978164,
+ "grad_norm": 0.0058724163100123405,
+ "learning_rate": 0.00011114617896724026,
+ "loss": 3.638444185256958,
+ "step": 3231
+ },
+ {
+ "epoch": 44.8943231441048,
+ "grad_norm": 0.006073266267776489,
+ "learning_rate": 0.00011071215563615752,
+ "loss": 3.680121421813965,
+ "step": 3232
+ },
+ {
+ "epoch": 44.90829694323144,
+ "grad_norm": 0.005904203746467829,
+ "learning_rate": 0.00011027926371442662,
+ "loss": 3.6883950233459473,
+ "step": 3233
+ },
+ {
+ "epoch": 44.92227074235808,
+ "grad_norm": 0.005889971274882555,
+ "learning_rate": 0.00010984750355082045,
+ "loss": 3.6875810623168945,
+ "step": 3234
+ },
+ {
+ "epoch": 44.93624454148472,
+ "grad_norm": 0.005878197029232979,
+ "learning_rate": 0.00010941687549320068,
+ "loss": 3.6341497898101807,
+ "step": 3235
+ },
+ {
+ "epoch": 44.95021834061135,
+ "grad_norm": 0.005862570833414793,
+ "learning_rate": 0.00010898737988851677,
+ "loss": 3.648092746734619,
+ "step": 3236
+ },
+ {
+ "epoch": 44.96419213973799,
+ "grad_norm": 0.0058378661051392555,
+ "learning_rate": 0.00010855901708280482,
+ "loss": 3.627786636352539,
+ "step": 3237
+ },
+ {
+ "epoch": 44.97816593886463,
+ "grad_norm": 0.0057954080402851105,
+ "learning_rate": 0.0001081317874211895,
+ "loss": 3.636265754699707,
+ "step": 3238
+ },
+ {
+ "epoch": 44.992139737991266,
+ "grad_norm": 0.005763230379670858,
+ "learning_rate": 0.00010770569124788201,
+ "loss": 3.6991324424743652,
+ "step": 3239
+ },
+ {
+ "epoch": 45.0,
+ "grad_norm": 0.007435750216245651,
+ "learning_rate": 0.00010728072890618029,
+ "loss": 3.698052406311035,
+ "step": 3240
+ },
+ {
+ "epoch": 45.0,
+ "eval_loss": 4.379567623138428,
+ "eval_runtime": 50.7833,
+ "eval_samples_per_second": 48.087,
+ "eval_steps_per_second": 1.516,
+ "step": 3240
+ },
+ {
+ "epoch": 45.01397379912664,
+ "grad_norm": 0.0059455581940710545,
+ "learning_rate": 0.00010685690073846878,
+ "loss": 3.649308204650879,
+ "step": 3241
+ },
+ {
+ "epoch": 45.02794759825328,
+ "grad_norm": 0.006021182518452406,
+ "learning_rate": 0.00010643420708621834,
+ "loss": 3.6339311599731445,
+ "step": 3242
+ },
+ {
+ "epoch": 45.041921397379916,
+ "grad_norm": 0.0061815716326236725,
+ "learning_rate": 0.00010601264828998569,
+ "loss": 3.658045530319214,
+ "step": 3243
+ },
+ {
+ "epoch": 45.05589519650655,
+ "grad_norm": 0.005834038369357586,
+ "learning_rate": 0.0001055922246894127,
+ "loss": 3.6551108360290527,
+ "step": 3244
+ },
+ {
+ "epoch": 45.069868995633186,
+ "grad_norm": 0.005903543904423714,
+ "learning_rate": 0.0001051729366232273,
+ "loss": 3.6415858268737793,
+ "step": 3245
+ },
+ {
+ "epoch": 45.083842794759825,
+ "grad_norm": 0.006033484358340502,
+ "learning_rate": 0.00010475478442924231,
+ "loss": 3.607917308807373,
+ "step": 3246
+ },
+ {
+ "epoch": 45.09781659388646,
+ "grad_norm": 0.005784410983324051,
+ "learning_rate": 0.00010433776844435538,
+ "loss": 3.641299247741699,
+ "step": 3247
+ },
+ {
+ "epoch": 45.1117903930131,
+ "grad_norm": 0.006221757270395756,
+ "learning_rate": 0.00010392188900454902,
+ "loss": 3.632112979888916,
+ "step": 3248
+ },
+ {
+ "epoch": 45.12576419213974,
+ "grad_norm": 0.005921157542616129,
+ "learning_rate": 0.00010350714644488925,
+ "loss": 3.6394639015197754,
+ "step": 3249
+ },
+ {
+ "epoch": 45.13973799126637,
+ "grad_norm": 0.005843059625476599,
+ "learning_rate": 0.00010309354109952728,
+ "loss": 3.6688497066497803,
+ "step": 3250
+ },
+ {
+ "epoch": 45.15371179039301,
+ "grad_norm": 0.006184449419379234,
+ "learning_rate": 0.000102681073301697,
+ "loss": 3.62900972366333,
+ "step": 3251
+ },
+ {
+ "epoch": 45.16768558951965,
+ "grad_norm": 0.005910454783588648,
+ "learning_rate": 0.00010226974338371635,
+ "loss": 3.639268398284912,
+ "step": 3252
+ },
+ {
+ "epoch": 45.18165938864629,
+ "grad_norm": 0.006613037548959255,
+ "learning_rate": 0.00010185955167698657,
+ "loss": 3.651017665863037,
+ "step": 3253
+ },
+ {
+ "epoch": 45.19563318777293,
+ "grad_norm": 0.006017698906362057,
+ "learning_rate": 0.00010145049851199155,
+ "loss": 3.6830077171325684,
+ "step": 3254
+ },
+ {
+ "epoch": 45.209606986899566,
+ "grad_norm": 0.006236966699361801,
+ "learning_rate": 0.0001010425842182984,
+ "loss": 3.6856958866119385,
+ "step": 3255
+ },
+ {
+ "epoch": 45.223580786026204,
+ "grad_norm": 0.0062433164566755295,
+ "learning_rate": 0.0001006358091245557,
+ "loss": 3.7161612510681152,
+ "step": 3256
+ },
+ {
+ "epoch": 45.237554585152836,
+ "grad_norm": 0.006063258275389671,
+ "learning_rate": 0.00010023017355849525,
+ "loss": 3.649462938308716,
+ "step": 3257
+ },
+ {
+ "epoch": 45.251528384279474,
+ "grad_norm": 0.006399665027856827,
+ "learning_rate": 9.982567784693017e-05,
+ "loss": 3.624393939971924,
+ "step": 3258
+ },
+ {
+ "epoch": 45.26550218340611,
+ "grad_norm": 0.006117025390267372,
+ "learning_rate": 9.94223223157554e-05,
+ "loss": 3.6675968170166016,
+ "step": 3259
+ },
+ {
+ "epoch": 45.27947598253275,
+ "grad_norm": 0.006158605683594942,
+ "learning_rate": 9.902010728994665e-05,
+ "loss": 3.6349360942840576,
+ "step": 3260
+ },
+ {
+ "epoch": 45.29344978165939,
+ "grad_norm": 0.0061307125724852085,
+ "learning_rate": 9.861903309356244e-05,
+ "loss": 3.6786162853240967,
+ "step": 3261
+ },
+ {
+ "epoch": 45.30742358078603,
+ "grad_norm": 0.005947567522525787,
+ "learning_rate": 9.821910004973994e-05,
+ "loss": 3.651076316833496,
+ "step": 3262
+ },
+ {
+ "epoch": 45.32139737991266,
+ "grad_norm": 0.0059933350421488285,
+ "learning_rate": 9.782030848069845e-05,
+ "loss": 3.6713902950286865,
+ "step": 3263
+ },
+ {
+ "epoch": 45.3353711790393,
+ "grad_norm": 0.006069707218557596,
+ "learning_rate": 9.74226587077372e-05,
+ "loss": 3.6770577430725098,
+ "step": 3264
+ },
+ {
+ "epoch": 45.34934497816594,
+ "grad_norm": 0.005641316995024681,
+ "learning_rate": 9.702615105123522e-05,
+ "loss": 3.6663057804107666,
+ "step": 3265
+ },
+ {
+ "epoch": 45.36331877729258,
+ "grad_norm": 0.006138736382126808,
+ "learning_rate": 9.663078583065183e-05,
+ "loss": 3.7058024406433105,
+ "step": 3266
+ },
+ {
+ "epoch": 45.377292576419215,
+ "grad_norm": 0.006074203178286552,
+ "learning_rate": 9.623656336452509e-05,
+ "loss": 3.684798002243042,
+ "step": 3267
+ },
+ {
+ "epoch": 45.391266375545854,
+ "grad_norm": 0.006269104778766632,
+ "learning_rate": 9.58434839704738e-05,
+ "loss": 3.653730630874634,
+ "step": 3268
+ },
+ {
+ "epoch": 45.40524017467249,
+ "grad_norm": 0.005954077932983637,
+ "learning_rate": 9.545154796519407e-05,
+ "loss": 3.631556272506714,
+ "step": 3269
+ },
+ {
+ "epoch": 45.419213973799124,
+ "grad_norm": 0.006305301561951637,
+ "learning_rate": 9.506075566446213e-05,
+ "loss": 3.652346134185791,
+ "step": 3270
+ },
+ {
+ "epoch": 45.43318777292576,
+ "grad_norm": 0.005923055112361908,
+ "learning_rate": 9.46711073831322e-05,
+ "loss": 3.644423007965088,
+ "step": 3271
+ },
+ {
+ "epoch": 45.4471615720524,
+ "grad_norm": 0.005840525031089783,
+ "learning_rate": 9.428260343513663e-05,
+ "loss": 3.664320468902588,
+ "step": 3272
+ },
+ {
+ "epoch": 45.46113537117904,
+ "grad_norm": 0.0058747464790940285,
+ "learning_rate": 9.389524413348661e-05,
+ "loss": 3.693209409713745,
+ "step": 3273
+ },
+ {
+ "epoch": 45.47510917030568,
+ "grad_norm": 0.005935903172940016,
+ "learning_rate": 9.350902979027003e-05,
+ "loss": 3.655853509902954,
+ "step": 3274
+ },
+ {
+ "epoch": 45.48908296943232,
+ "grad_norm": 0.005829546134918928,
+ "learning_rate": 9.312396071665273e-05,
+ "loss": 3.645937442779541,
+ "step": 3275
+ },
+ {
+ "epoch": 45.50305676855895,
+ "grad_norm": 0.005715377628803253,
+ "learning_rate": 9.274003722287829e-05,
+ "loss": 3.660888671875,
+ "step": 3276
+ },
+ {
+ "epoch": 45.51703056768559,
+ "grad_norm": 0.005914001259952784,
+ "learning_rate": 9.235725961826662e-05,
+ "loss": 3.6928553581237793,
+ "step": 3277
+ },
+ {
+ "epoch": 45.531004366812226,
+ "grad_norm": 0.006055662874132395,
+ "learning_rate": 9.197562821121495e-05,
+ "loss": 3.6630935668945312,
+ "step": 3278
+ },
+ {
+ "epoch": 45.544978165938865,
+ "grad_norm": 0.005942841526120901,
+ "learning_rate": 9.159514330919671e-05,
+ "loss": 3.6434600353240967,
+ "step": 3279
+ },
+ {
+ "epoch": 45.5589519650655,
+ "grad_norm": 0.005879723932594061,
+ "learning_rate": 9.121580521876192e-05,
+ "loss": 3.6828930377960205,
+ "step": 3280
+ },
+ {
+ "epoch": 45.57292576419214,
+ "grad_norm": 0.005810475442558527,
+ "learning_rate": 9.083761424553615e-05,
+ "loss": 3.6644086837768555,
+ "step": 3281
+ },
+ {
+ "epoch": 45.58689956331878,
+ "grad_norm": 0.0058500380255281925,
+ "learning_rate": 9.046057069422091e-05,
+ "loss": 3.6613426208496094,
+ "step": 3282
+ },
+ {
+ "epoch": 45.60087336244541,
+ "grad_norm": 0.005881919525563717,
+ "learning_rate": 9.008467486859382e-05,
+ "loss": 3.646371841430664,
+ "step": 3283
+ },
+ {
+ "epoch": 45.61484716157205,
+ "grad_norm": 0.00608665868639946,
+ "learning_rate": 8.970992707150702e-05,
+ "loss": 3.6912505626678467,
+ "step": 3284
+ },
+ {
+ "epoch": 45.62882096069869,
+ "grad_norm": 0.0059072356671094894,
+ "learning_rate": 8.933632760488831e-05,
+ "loss": 3.6508445739746094,
+ "step": 3285
+ },
+ {
+ "epoch": 45.64279475982533,
+ "grad_norm": 0.00589236943051219,
+ "learning_rate": 8.896387676973956e-05,
+ "loss": 3.636539936065674,
+ "step": 3286
+ },
+ {
+ "epoch": 45.65676855895197,
+ "grad_norm": 0.006131782662123442,
+ "learning_rate": 8.859257486613818e-05,
+ "loss": 3.643298625946045,
+ "step": 3287
+ },
+ {
+ "epoch": 45.670742358078606,
+ "grad_norm": 0.006039046682417393,
+ "learning_rate": 8.82224221932353e-05,
+ "loss": 3.6826415061950684,
+ "step": 3288
+ },
+ {
+ "epoch": 45.68471615720524,
+ "grad_norm": 0.005942283198237419,
+ "learning_rate": 8.785341904925602e-05,
+ "loss": 3.6603708267211914,
+ "step": 3289
+ },
+ {
+ "epoch": 45.698689956331876,
+ "grad_norm": 0.006028817500919104,
+ "learning_rate": 8.748556573149962e-05,
+ "loss": 3.6836555004119873,
+ "step": 3290
+ },
+ {
+ "epoch": 45.712663755458514,
+ "grad_norm": 0.006245827302336693,
+ "learning_rate": 8.711886253633896e-05,
+ "loss": 3.689937114715576,
+ "step": 3291
+ },
+ {
+ "epoch": 45.72663755458515,
+ "grad_norm": 0.005829308647662401,
+ "learning_rate": 8.675330975922034e-05,
+ "loss": 3.646091938018799,
+ "step": 3292
+ },
+ {
+ "epoch": 45.74061135371179,
+ "grad_norm": 0.005799743812531233,
+ "learning_rate": 8.63889076946624e-05,
+ "loss": 3.667741298675537,
+ "step": 3293
+ },
+ {
+ "epoch": 45.75458515283843,
+ "grad_norm": 0.006182873621582985,
+ "learning_rate": 8.602565663625812e-05,
+ "loss": 3.6704702377319336,
+ "step": 3294
+ },
+ {
+ "epoch": 45.76855895196506,
+ "grad_norm": 0.0059661525301635265,
+ "learning_rate": 8.566355687667166e-05,
+ "loss": 3.6082215309143066,
+ "step": 3295
+ },
+ {
+ "epoch": 45.7825327510917,
+ "grad_norm": 0.006096175871789455,
+ "learning_rate": 8.530260870764065e-05,
+ "loss": 3.6541519165039062,
+ "step": 3296
+ },
+ {
+ "epoch": 45.79650655021834,
+ "grad_norm": 0.005957531277090311,
+ "learning_rate": 8.494281241997411e-05,
+ "loss": 3.6488404273986816,
+ "step": 3297
+ },
+ {
+ "epoch": 45.81048034934498,
+ "grad_norm": 0.006362909451127052,
+ "learning_rate": 8.458416830355393e-05,
+ "loss": 3.6703639030456543,
+ "step": 3298
+ },
+ {
+ "epoch": 45.82445414847162,
+ "grad_norm": 0.005813185591250658,
+ "learning_rate": 8.422667664733304e-05,
+ "loss": 3.6723361015319824,
+ "step": 3299
+ },
+ {
+ "epoch": 45.838427947598255,
+ "grad_norm": 0.005782100837677717,
+ "learning_rate": 8.387033773933557e-05,
+ "loss": 3.63480806350708,
+ "step": 3300
+ },
+ {
+ "epoch": 45.852401746724894,
+ "grad_norm": 0.006054454483091831,
+ "learning_rate": 8.351515186665794e-05,
+ "loss": 3.6608922481536865,
+ "step": 3301
+ },
+ {
+ "epoch": 45.866375545851525,
+ "grad_norm": 0.006052988115698099,
+ "learning_rate": 8.316111931546666e-05,
+ "loss": 3.6687750816345215,
+ "step": 3302
+ },
+ {
+ "epoch": 45.880349344978164,
+ "grad_norm": 0.006084649357944727,
+ "learning_rate": 8.280824037099941e-05,
+ "loss": 3.6396217346191406,
+ "step": 3303
+ },
+ {
+ "epoch": 45.8943231441048,
+ "grad_norm": 0.005772392265498638,
+ "learning_rate": 8.245651531756433e-05,
+ "loss": 3.7278900146484375,
+ "step": 3304
+ },
+ {
+ "epoch": 45.90829694323144,
+ "grad_norm": 0.005887683946639299,
+ "learning_rate": 8.210594443854007e-05,
+ "loss": 3.6623833179473877,
+ "step": 3305
+ },
+ {
+ "epoch": 45.92227074235808,
+ "grad_norm": 0.005802373867481947,
+ "learning_rate": 8.175652801637535e-05,
+ "loss": 3.680635690689087,
+ "step": 3306
+ },
+ {
+ "epoch": 45.93624454148472,
+ "grad_norm": 0.005748553667217493,
+ "learning_rate": 8.14082663325883e-05,
+ "loss": 3.6317763328552246,
+ "step": 3307
+ },
+ {
+ "epoch": 45.95021834061135,
+ "grad_norm": 0.005717726889997721,
+ "learning_rate": 8.106115966776743e-05,
+ "loss": 3.6421971321105957,
+ "step": 3308
+ },
+ {
+ "epoch": 45.96419213973799,
+ "grad_norm": 0.005919075571000576,
+ "learning_rate": 8.071520830156999e-05,
+ "loss": 3.726754665374756,
+ "step": 3309
+ },
+ {
+ "epoch": 45.97816593886463,
+ "grad_norm": 0.005714495666325092,
+ "learning_rate": 8.037041251272334e-05,
+ "loss": 3.637263298034668,
+ "step": 3310
+ },
+ {
+ "epoch": 45.992139737991266,
+ "grad_norm": 0.005621272139251232,
+ "learning_rate": 8.002677257902252e-05,
+ "loss": 3.6489784717559814,
+ "step": 3311
+ },
+ {
+ "epoch": 46.0,
+ "grad_norm": 0.007174668367952108,
+ "learning_rate": 7.968428877733201e-05,
+ "loss": 3.67305326461792,
+ "step": 3312
+ },
+ {
+ "epoch": 46.0,
+ "eval_loss": 4.380944728851318,
+ "eval_runtime": 51.6456,
+ "eval_samples_per_second": 47.284,
+ "eval_steps_per_second": 1.491,
+ "step": 3312
+ },
+ {
+ "epoch": 46.01397379912664,
+ "grad_norm": 0.00578242726624012,
+ "learning_rate": 7.934296138358568e-05,
+ "loss": 3.6154122352600098,
+ "step": 3313
+ },
+ {
+ "epoch": 46.02794759825328,
+ "grad_norm": 0.005650710314512253,
+ "learning_rate": 7.900279067278403e-05,
+ "loss": 3.6636948585510254,
+ "step": 3314
+ },
+ {
+ "epoch": 46.041921397379916,
+ "grad_norm": 0.0056931874714791775,
+ "learning_rate": 7.866377691899657e-05,
+ "loss": 3.636855125427246,
+ "step": 3315
+ },
+ {
+ "epoch": 46.05589519650655,
+ "grad_norm": 0.005795546341687441,
+ "learning_rate": 7.832592039536082e-05,
+ "loss": 3.6582765579223633,
+ "step": 3316
+ },
+ {
+ "epoch": 46.069868995633186,
+ "grad_norm": 0.005859258119016886,
+ "learning_rate": 7.798922137408176e-05,
+ "loss": 3.6686697006225586,
+ "step": 3317
+ },
+ {
+ "epoch": 46.083842794759825,
+ "grad_norm": 0.005823366343975067,
+ "learning_rate": 7.765368012643131e-05,
+ "loss": 3.66125750541687,
+ "step": 3318
+ },
+ {
+ "epoch": 46.09781659388646,
+ "grad_norm": 0.005628711078315973,
+ "learning_rate": 7.73192969227491e-05,
+ "loss": 3.65299129486084,
+ "step": 3319
+ },
+ {
+ "epoch": 46.1117903930131,
+ "grad_norm": 0.005935952067375183,
+ "learning_rate": 7.698607203244189e-05,
+ "loss": 3.63519287109375,
+ "step": 3320
+ },
+ {
+ "epoch": 46.12576419213974,
+ "grad_norm": 0.005684727802872658,
+ "learning_rate": 7.665400572398285e-05,
+ "loss": 3.6002516746520996,
+ "step": 3321
+ },
+ {
+ "epoch": 46.13973799126637,
+ "grad_norm": 0.005795048549771309,
+ "learning_rate": 7.632309826491177e-05,
+ "loss": 3.6974287033081055,
+ "step": 3322
+ },
+ {
+ "epoch": 46.15371179039301,
+ "grad_norm": 0.0059766992926597595,
+ "learning_rate": 7.599334992183438e-05,
+ "loss": 3.672163486480713,
+ "step": 3323
+ },
+ {
+ "epoch": 46.16768558951965,
+ "grad_norm": 0.005643186625093222,
+ "learning_rate": 7.566476096042397e-05,
+ "loss": 3.6390175819396973,
+ "step": 3324
+ },
+ {
+ "epoch": 46.18165938864629,
+ "grad_norm": 0.0058948020450770855,
+ "learning_rate": 7.533733164541767e-05,
+ "loss": 3.635690689086914,
+ "step": 3325
+ },
+ {
+ "epoch": 46.19563318777293,
+ "grad_norm": 0.005942242220044136,
+ "learning_rate": 7.501106224062e-05,
+ "loss": 3.6397767066955566,
+ "step": 3326
+ },
+ {
+ "epoch": 46.209606986899566,
+ "grad_norm": 0.0054268259555101395,
+ "learning_rate": 7.468595300890026e-05,
+ "loss": 3.634727954864502,
+ "step": 3327
+ },
+ {
+ "epoch": 46.223580786026204,
+ "grad_norm": 0.006279955618083477,
+ "learning_rate": 7.43620042121929e-05,
+ "loss": 3.6551690101623535,
+ "step": 3328
+ },
+ {
+ "epoch": 46.237554585152836,
+ "grad_norm": 0.005737216677516699,
+ "learning_rate": 7.403921611149778e-05,
+ "loss": 3.665435314178467,
+ "step": 3329
+ },
+ {
+ "epoch": 46.251528384279474,
+ "grad_norm": 0.005809346679598093,
+ "learning_rate": 7.371758896687908e-05,
+ "loss": 3.645390033721924,
+ "step": 3330
+ },
+ {
+ "epoch": 46.26550218340611,
+ "grad_norm": 0.005774138495326042,
+ "learning_rate": 7.339712303746658e-05,
+ "loss": 3.6289312839508057,
+ "step": 3331
+ },
+ {
+ "epoch": 46.27947598253275,
+ "grad_norm": 0.005683782510459423,
+ "learning_rate": 7.30778185814535e-05,
+ "loss": 3.6434290409088135,
+ "step": 3332
+ },
+ {
+ "epoch": 46.29344978165939,
+ "grad_norm": 0.006072668358683586,
+ "learning_rate": 7.275967585609755e-05,
+ "loss": 3.6633148193359375,
+ "step": 3333
+ },
+ {
+ "epoch": 46.30742358078603,
+ "grad_norm": 0.005564856808632612,
+ "learning_rate": 7.244269511772106e-05,
+ "loss": 3.617435932159424,
+ "step": 3334
+ },
+ {
+ "epoch": 46.32139737991266,
+ "grad_norm": 0.005909780040383339,
+ "learning_rate": 7.212687662170915e-05,
+ "loss": 3.6613821983337402,
+ "step": 3335
+ },
+ {
+ "epoch": 46.3353711790393,
+ "grad_norm": 0.005920986644923687,
+ "learning_rate": 7.181222062251155e-05,
+ "loss": 3.6254312992095947,
+ "step": 3336
+ },
+ {
+ "epoch": 46.34934497816594,
+ "grad_norm": 0.005616472102701664,
+ "learning_rate": 7.14987273736407e-05,
+ "loss": 3.653961658477783,
+ "step": 3337
+ },
+ {
+ "epoch": 46.36331877729258,
+ "grad_norm": 0.006085917819291353,
+ "learning_rate": 7.118639712767222e-05,
+ "loss": 3.6716418266296387,
+ "step": 3338
+ },
+ {
+ "epoch": 46.377292576419215,
+ "grad_norm": 0.0057594352401793,
+ "learning_rate": 7.087523013624534e-05,
+ "loss": 3.66914701461792,
+ "step": 3339
+ },
+ {
+ "epoch": 46.391266375545854,
+ "grad_norm": 0.005555839743465185,
+ "learning_rate": 7.056522665006162e-05,
+ "loss": 3.674562692642212,
+ "step": 3340
+ },
+ {
+ "epoch": 46.40524017467249,
+ "grad_norm": 0.005709810648113489,
+ "learning_rate": 7.025638691888476e-05,
+ "loss": 3.664320468902588,
+ "step": 3341
+ },
+ {
+ "epoch": 46.419213973799124,
+ "grad_norm": 0.005785972345620394,
+ "learning_rate": 6.994871119154197e-05,
+ "loss": 3.6416711807250977,
+ "step": 3342
+ },
+ {
+ "epoch": 46.43318777292576,
+ "grad_norm": 0.0058121830224990845,
+ "learning_rate": 6.964219971592198e-05,
+ "loss": 3.654954433441162,
+ "step": 3343
+ },
+ {
+ "epoch": 46.4471615720524,
+ "grad_norm": 0.005938741844147444,
+ "learning_rate": 6.933685273897533e-05,
+ "loss": 3.6046786308288574,
+ "step": 3344
+ },
+ {
+ "epoch": 46.46113537117904,
+ "grad_norm": 0.005746894516050816,
+ "learning_rate": 6.903267050671447e-05,
+ "loss": 3.6525707244873047,
+ "step": 3345
+ },
+ {
+ "epoch": 46.47510917030568,
+ "grad_norm": 0.005644070915877819,
+ "learning_rate": 6.872965326421365e-05,
+ "loss": 3.652769088745117,
+ "step": 3346
+ },
+ {
+ "epoch": 46.48908296943232,
+ "grad_norm": 0.006110438145697117,
+ "learning_rate": 6.84278012556088e-05,
+ "loss": 3.669772148132324,
+ "step": 3347
+ },
+ {
+ "epoch": 46.50305676855895,
+ "grad_norm": 0.0055745323188602924,
+ "learning_rate": 6.81271147240959e-05,
+ "loss": 3.6494526863098145,
+ "step": 3348
+ },
+ {
+ "epoch": 46.51703056768559,
+ "grad_norm": 0.005900663323700428,
+ "learning_rate": 6.78275939119328e-05,
+ "loss": 3.6548643112182617,
+ "step": 3349
+ },
+ {
+ "epoch": 46.531004366812226,
+ "grad_norm": 0.00568796694278717,
+ "learning_rate": 6.752923906043862e-05,
+ "loss": 3.6537368297576904,
+ "step": 3350
+ },
+ {
+ "epoch": 46.544978165938865,
+ "grad_norm": 0.005612431094050407,
+ "learning_rate": 6.723205040999193e-05,
+ "loss": 3.6221089363098145,
+ "step": 3351
+ },
+ {
+ "epoch": 46.5589519650655,
+ "grad_norm": 0.005935634020715952,
+ "learning_rate": 6.693602820003248e-05,
+ "loss": 3.6739964485168457,
+ "step": 3352
+ },
+ {
+ "epoch": 46.57292576419214,
+ "grad_norm": 0.005767039954662323,
+ "learning_rate": 6.66411726690598e-05,
+ "loss": 3.6298999786376953,
+ "step": 3353
+ },
+ {
+ "epoch": 46.58689956331878,
+ "grad_norm": 0.005834923591464758,
+ "learning_rate": 6.634748405463404e-05,
+ "loss": 3.6653504371643066,
+ "step": 3354
+ },
+ {
+ "epoch": 46.60087336244541,
+ "grad_norm": 0.005940297152847052,
+ "learning_rate": 6.605496259337425e-05,
+ "loss": 3.6638693809509277,
+ "step": 3355
+ },
+ {
+ "epoch": 46.61484716157205,
+ "grad_norm": 0.00571000250056386,
+ "learning_rate": 6.576360852096006e-05,
+ "loss": 3.639021873474121,
+ "step": 3356
+ },
+ {
+ "epoch": 46.62882096069869,
+ "grad_norm": 0.0057969787158071995,
+ "learning_rate": 6.547342207213038e-05,
+ "loss": 3.6597180366516113,
+ "step": 3357
+ },
+ {
+ "epoch": 46.64279475982533,
+ "grad_norm": 0.005795357748866081,
+ "learning_rate": 6.518440348068267e-05,
+ "loss": 3.6444644927978516,
+ "step": 3358
+ },
+ {
+ "epoch": 46.65676855895197,
+ "grad_norm": 0.005731288809329271,
+ "learning_rate": 6.489655297947441e-05,
+ "loss": 3.668120861053467,
+ "step": 3359
+ },
+ {
+ "epoch": 46.670742358078606,
+ "grad_norm": 0.0058074090629816055,
+ "learning_rate": 6.460987080042126e-05,
+ "loss": 3.639407157897949,
+ "step": 3360
+ },
+ {
+ "epoch": 46.68471615720524,
+ "grad_norm": 0.005735141225159168,
+ "learning_rate": 6.432435717449837e-05,
+ "loss": 3.6410751342773438,
+ "step": 3361
+ },
+ {
+ "epoch": 46.698689956331876,
+ "grad_norm": 0.005885255988687277,
+ "learning_rate": 6.404001233173828e-05,
+ "loss": 3.658564567565918,
+ "step": 3362
+ },
+ {
+ "epoch": 46.712663755458514,
+ "grad_norm": 0.005977948661893606,
+ "learning_rate": 6.375683650123268e-05,
+ "loss": 3.688422918319702,
+ "step": 3363
+ },
+ {
+ "epoch": 46.72663755458515,
+ "grad_norm": 0.005629570223391056,
+ "learning_rate": 6.347482991113151e-05,
+ "loss": 3.6477694511413574,
+ "step": 3364
+ },
+ {
+ "epoch": 46.74061135371179,
+ "grad_norm": 0.005695016123354435,
+ "learning_rate": 6.319399278864202e-05,
+ "loss": 3.6869497299194336,
+ "step": 3365
+ },
+ {
+ "epoch": 46.75458515283843,
+ "grad_norm": 0.0057486798614263535,
+ "learning_rate": 6.291432536003e-05,
+ "loss": 3.6609716415405273,
+ "step": 3366
+ },
+ {
+ "epoch": 46.76855895196506,
+ "grad_norm": 0.005637758877128363,
+ "learning_rate": 6.263582785061794e-05,
+ "loss": 3.644331932067871,
+ "step": 3367
+ },
+ {
+ "epoch": 46.7825327510917,
+ "grad_norm": 0.005615375470370054,
+ "learning_rate": 6.235850048478689e-05,
+ "loss": 3.6516261100769043,
+ "step": 3368
+ },
+ {
+ "epoch": 46.79650655021834,
+ "grad_norm": 0.005714157596230507,
+ "learning_rate": 6.208234348597384e-05,
+ "loss": 3.6313061714172363,
+ "step": 3369
+ },
+ {
+ "epoch": 46.81048034934498,
+ "grad_norm": 0.005742870736867189,
+ "learning_rate": 6.180735707667399e-05,
+ "loss": 3.658287525177002,
+ "step": 3370
+ },
+ {
+ "epoch": 46.82445414847162,
+ "grad_norm": 0.005571036599576473,
+ "learning_rate": 6.153354147843854e-05,
+ "loss": 3.6537227630615234,
+ "step": 3371
+ },
+ {
+ "epoch": 46.838427947598255,
+ "grad_norm": 0.005676159635186195,
+ "learning_rate": 6.126089691187606e-05,
+ "loss": 3.6534817218780518,
+ "step": 3372
+ },
+ {
+ "epoch": 46.852401746724894,
+ "grad_norm": 0.00567635940387845,
+ "learning_rate": 6.098942359665141e-05,
+ "loss": 3.62833309173584,
+ "step": 3373
+ },
+ {
+ "epoch": 46.866375545851525,
+ "grad_norm": 0.005548842716962099,
+ "learning_rate": 6.071912175148553e-05,
+ "loss": 3.6426186561584473,
+ "step": 3374
+ },
+ {
+ "epoch": 46.880349344978164,
+ "grad_norm": 0.0056854793801903725,
+ "learning_rate": 6.0449991594155694e-05,
+ "loss": 3.670558452606201,
+ "step": 3375
+ },
+ {
+ "epoch": 46.8943231441048,
+ "grad_norm": 0.005691561382263899,
+ "learning_rate": 6.0182033341495317e-05,
+ "loss": 3.6603145599365234,
+ "step": 3376
+ },
+ {
+ "epoch": 46.90829694323144,
+ "grad_norm": 0.005610436201095581,
+ "learning_rate": 5.991524720939341e-05,
+ "loss": 3.6651010513305664,
+ "step": 3377
+ },
+ {
+ "epoch": 46.92227074235808,
+ "grad_norm": 0.005553451832383871,
+ "learning_rate": 5.9649633412794984e-05,
+ "loss": 3.680150032043457,
+ "step": 3378
+ },
+ {
+ "epoch": 46.93624454148472,
+ "grad_norm": 0.005749273579567671,
+ "learning_rate": 5.93851921657e-05,
+ "loss": 3.6609292030334473,
+ "step": 3379
+ },
+ {
+ "epoch": 46.95021834061135,
+ "grad_norm": 0.005574697162955999,
+ "learning_rate": 5.91219236811644e-05,
+ "loss": 3.7022757530212402,
+ "step": 3380
+ },
+ {
+ "epoch": 46.96419213973799,
+ "grad_norm": 0.005650301463901997,
+ "learning_rate": 5.885982817129827e-05,
+ "loss": 3.71413516998291,
+ "step": 3381
+ },
+ {
+ "epoch": 46.97816593886463,
+ "grad_norm": 0.005744843743741512,
+ "learning_rate": 5.859890584726769e-05,
+ "loss": 3.6550443172454834,
+ "step": 3382
+ },
+ {
+ "epoch": 46.992139737991266,
+ "grad_norm": 0.00542390625923872,
+ "learning_rate": 5.833915691929282e-05,
+ "loss": 3.6864724159240723,
+ "step": 3383
+ },
+ {
+ "epoch": 47.0,
+ "grad_norm": 0.007102064788341522,
+ "learning_rate": 5.808058159664905e-05,
+ "loss": 3.681469440460205,
+ "step": 3384
+ },
+ {
+ "epoch": 47.0,
+ "eval_loss": 4.382168769836426,
+ "eval_runtime": 52.2293,
+ "eval_samples_per_second": 46.755,
+ "eval_steps_per_second": 1.474,
+ "step": 3384
+ },
+ {
+ "epoch": 47.01397379912664,
+ "grad_norm": 0.005845201667398214,
+ "learning_rate": 5.782318008766519e-05,
+ "loss": 3.63492751121521,
+ "step": 3385
+ },
+ {
+ "epoch": 47.02794759825328,
+ "grad_norm": 0.005614436697214842,
+ "learning_rate": 5.75669525997254e-05,
+ "loss": 3.629580020904541,
+ "step": 3386
+ },
+ {
+ "epoch": 47.041921397379916,
+ "grad_norm": 0.0055298879742622375,
+ "learning_rate": 5.731189933926781e-05,
+ "loss": 3.6301462650299072,
+ "step": 3387
+ },
+ {
+ "epoch": 47.05589519650655,
+ "grad_norm": 0.005644683260470629,
+ "learning_rate": 5.7058020511783706e-05,
+ "loss": 3.6150715351104736,
+ "step": 3388
+ },
+ {
+ "epoch": 47.069868995633186,
+ "grad_norm": 0.005629776511341333,
+ "learning_rate": 5.6805316321818886e-05,
+ "loss": 3.6537346839904785,
+ "step": 3389
+ },
+ {
+ "epoch": 47.083842794759825,
+ "grad_norm": 0.005810001865029335,
+ "learning_rate": 5.655378697297258e-05,
+ "loss": 3.64941143989563,
+ "step": 3390
+ },
+ {
+ "epoch": 47.09781659388646,
+ "grad_norm": 0.0057138605043292046,
+ "learning_rate": 5.630343266789741e-05,
+ "loss": 3.6648316383361816,
+ "step": 3391
+ },
+ {
+ "epoch": 47.1117903930131,
+ "grad_norm": 0.005495130084455013,
+ "learning_rate": 5.605425360829931e-05,
+ "loss": 3.619474411010742,
+ "step": 3392
+ },
+ {
+ "epoch": 47.12576419213974,
+ "grad_norm": 0.005678716115653515,
+ "learning_rate": 5.580624999493715e-05,
+ "loss": 3.6268370151519775,
+ "step": 3393
+ },
+ {
+ "epoch": 47.13973799126637,
+ "grad_norm": 0.0057465555146336555,
+ "learning_rate": 5.5559422027623556e-05,
+ "loss": 3.6618542671203613,
+ "step": 3394
+ },
+ {
+ "epoch": 47.15371179039301,
+ "grad_norm": 0.005927131045609713,
+ "learning_rate": 5.5313769905222756e-05,
+ "loss": 3.6629247665405273,
+ "step": 3395
+ },
+ {
+ "epoch": 47.16768558951965,
+ "grad_norm": 0.0056359837763011456,
+ "learning_rate": 5.506929382565235e-05,
+ "loss": 3.638967990875244,
+ "step": 3396
+ },
+ {
+ "epoch": 47.18165938864629,
+ "grad_norm": 0.005787394475191832,
+ "learning_rate": 5.482599398588257e-05,
+ "loss": 3.6664178371429443,
+ "step": 3397
+ },
+ {
+ "epoch": 47.19563318777293,
+ "grad_norm": 0.0054801893420517445,
+ "learning_rate": 5.458387058193552e-05,
+ "loss": 3.6431665420532227,
+ "step": 3398
+ },
+ {
+ "epoch": 47.209606986899566,
+ "grad_norm": 0.005605428479611874,
+ "learning_rate": 5.434292380888563e-05,
+ "loss": 3.6501848697662354,
+ "step": 3399
+ },
+ {
+ "epoch": 47.223580786026204,
+ "grad_norm": 0.005695626139640808,
+ "learning_rate": 5.410315386085934e-05,
+ "loss": 3.6499099731445312,
+ "step": 3400
+ },
+ {
+ "epoch": 47.237554585152836,
+ "grad_norm": 0.005612112116068602,
+ "learning_rate": 5.386456093103495e-05,
+ "loss": 3.6890406608581543,
+ "step": 3401
+ },
+ {
+ "epoch": 47.251528384279474,
+ "grad_norm": 0.005486638285219669,
+ "learning_rate": 5.362714521164272e-05,
+ "loss": 3.6657395362854004,
+ "step": 3402
+ },
+ {
+ "epoch": 47.26550218340611,
+ "grad_norm": 0.005863480269908905,
+ "learning_rate": 5.3390906893963974e-05,
+ "loss": 3.666640281677246,
+ "step": 3403
+ },
+ {
+ "epoch": 47.27947598253275,
+ "grad_norm": 0.00578767666593194,
+ "learning_rate": 5.315584616833154e-05,
+ "loss": 3.6408395767211914,
+ "step": 3404
+ },
+ {
+ "epoch": 47.29344978165939,
+ "grad_norm": 0.005513481795787811,
+ "learning_rate": 5.2921963224130086e-05,
+ "loss": 3.661557197570801,
+ "step": 3405
+ },
+ {
+ "epoch": 47.30742358078603,
+ "grad_norm": 0.0056565022096037865,
+ "learning_rate": 5.268925824979431e-05,
+ "loss": 3.646548271179199,
+ "step": 3406
+ },
+ {
+ "epoch": 47.32139737991266,
+ "grad_norm": 0.0057138074189424515,
+ "learning_rate": 5.2457731432810745e-05,
+ "loss": 3.654798984527588,
+ "step": 3407
+ },
+ {
+ "epoch": 47.3353711790393,
+ "grad_norm": 0.005857877433300018,
+ "learning_rate": 5.222738295971629e-05,
+ "loss": 3.6370182037353516,
+ "step": 3408
+ },
+ {
+ "epoch": 47.34934497816594,
+ "grad_norm": 0.005622264463454485,
+ "learning_rate": 5.199821301609856e-05,
+ "loss": 3.6596436500549316,
+ "step": 3409
+ },
+ {
+ "epoch": 47.36331877729258,
+ "grad_norm": 0.005421077832579613,
+ "learning_rate": 5.1770221786595654e-05,
+ "loss": 3.638583183288574,
+ "step": 3410
+ },
+ {
+ "epoch": 47.377292576419215,
+ "grad_norm": 0.0054649352096021175,
+ "learning_rate": 5.154340945489587e-05,
+ "loss": 3.6493330001831055,
+ "step": 3411
+ },
+ {
+ "epoch": 47.391266375545854,
+ "grad_norm": 0.005776858888566494,
+ "learning_rate": 5.1317776203737905e-05,
+ "loss": 3.6507070064544678,
+ "step": 3412
+ },
+ {
+ "epoch": 47.40524017467249,
+ "grad_norm": 0.005619521252810955,
+ "learning_rate": 5.109332221491043e-05,
+ "loss": 3.635849952697754,
+ "step": 3413
+ },
+ {
+ "epoch": 47.419213973799124,
+ "grad_norm": 0.0055708554573357105,
+ "learning_rate": 5.0870047669251846e-05,
+ "loss": 3.68941068649292,
+ "step": 3414
+ },
+ {
+ "epoch": 47.43318777292576,
+ "grad_norm": 0.005647908430546522,
+ "learning_rate": 5.06479527466506e-05,
+ "loss": 3.666015625,
+ "step": 3415
+ },
+ {
+ "epoch": 47.4471615720524,
+ "grad_norm": 0.005934152286499739,
+ "learning_rate": 5.0427037626044326e-05,
+ "loss": 3.6556167602539062,
+ "step": 3416
+ },
+ {
+ "epoch": 47.46113537117904,
+ "grad_norm": 0.005791633389890194,
+ "learning_rate": 5.020730248542068e-05,
+ "loss": 3.6541948318481445,
+ "step": 3417
+ },
+ {
+ "epoch": 47.47510917030568,
+ "grad_norm": 0.00573629979044199,
+ "learning_rate": 4.998874750181586e-05,
+ "loss": 3.626328468322754,
+ "step": 3418
+ },
+ {
+ "epoch": 47.48908296943232,
+ "grad_norm": 0.005624848417937756,
+ "learning_rate": 4.9771372851315965e-05,
+ "loss": 3.6370186805725098,
+ "step": 3419
+ },
+ {
+ "epoch": 47.50305676855895,
+ "grad_norm": 0.005743588786572218,
+ "learning_rate": 4.955517870905577e-05,
+ "loss": 3.647320032119751,
+ "step": 3420
+ },
+ {
+ "epoch": 47.51703056768559,
+ "grad_norm": 0.005720328539609909,
+ "learning_rate": 4.934016524921894e-05,
+ "loss": 3.607609272003174,
+ "step": 3421
+ },
+ {
+ "epoch": 47.531004366812226,
+ "grad_norm": 0.005619847681373358,
+ "learning_rate": 4.912633264503801e-05,
+ "loss": 3.676149845123291,
+ "step": 3422
+ },
+ {
+ "epoch": 47.544978165938865,
+ "grad_norm": 0.00549230445176363,
+ "learning_rate": 4.8913681068793916e-05,
+ "loss": 3.6969799995422363,
+ "step": 3423
+ },
+ {
+ "epoch": 47.5589519650655,
+ "grad_norm": 0.0055993059650063515,
+ "learning_rate": 4.8702210691816596e-05,
+ "loss": 3.6594631671905518,
+ "step": 3424
+ },
+ {
+ "epoch": 47.57292576419214,
+ "grad_norm": 0.005832983180880547,
+ "learning_rate": 4.84919216844834e-05,
+ "loss": 3.6583518981933594,
+ "step": 3425
+ },
+ {
+ "epoch": 47.58689956331878,
+ "grad_norm": 0.005687489174306393,
+ "learning_rate": 4.828281421622057e-05,
+ "loss": 3.6750564575195312,
+ "step": 3426
+ },
+ {
+ "epoch": 47.60087336244541,
+ "grad_norm": 0.0055601997300982475,
+ "learning_rate": 4.807488845550223e-05,
+ "loss": 3.6731209754943848,
+ "step": 3427
+ },
+ {
+ "epoch": 47.61484716157205,
+ "grad_norm": 0.005822640378028154,
+ "learning_rate": 4.786814456985036e-05,
+ "loss": 3.6643764972686768,
+ "step": 3428
+ },
+ {
+ "epoch": 47.62882096069869,
+ "grad_norm": 0.0057311588898301125,
+ "learning_rate": 4.7662582725835025e-05,
+ "loss": 3.6473302841186523,
+ "step": 3429
+ },
+ {
+ "epoch": 47.64279475982533,
+ "grad_norm": 0.005722011439502239,
+ "learning_rate": 4.745820308907318e-05,
+ "loss": 3.6761927604675293,
+ "step": 3430
+ },
+ {
+ "epoch": 47.65676855895197,
+ "grad_norm": 0.005630903877317905,
+ "learning_rate": 4.725500582423046e-05,
+ "loss": 3.668677806854248,
+ "step": 3431
+ },
+ {
+ "epoch": 47.670742358078606,
+ "grad_norm": 0.005585893988609314,
+ "learning_rate": 4.7052991095018625e-05,
+ "loss": 3.6321611404418945,
+ "step": 3432
+ },
+ {
+ "epoch": 47.68471615720524,
+ "grad_norm": 0.005564410239458084,
+ "learning_rate": 4.6852159064197785e-05,
+ "loss": 3.649759531021118,
+ "step": 3433
+ },
+ {
+ "epoch": 47.698689956331876,
+ "grad_norm": 0.005554834380745888,
+ "learning_rate": 4.665250989357429e-05,
+ "loss": 3.6571452617645264,
+ "step": 3434
+ },
+ {
+ "epoch": 47.712663755458514,
+ "grad_norm": 0.005567575339227915,
+ "learning_rate": 4.645404374400205e-05,
+ "loss": 3.653958320617676,
+ "step": 3435
+ },
+ {
+ "epoch": 47.72663755458515,
+ "grad_norm": 0.00546268792822957,
+ "learning_rate": 4.62567607753817e-05,
+ "loss": 3.6481666564941406,
+ "step": 3436
+ },
+ {
+ "epoch": 47.74061135371179,
+ "grad_norm": 0.0055222418159246445,
+ "learning_rate": 4.606066114666047e-05,
+ "loss": 3.6125011444091797,
+ "step": 3437
+ },
+ {
+ "epoch": 47.75458515283843,
+ "grad_norm": 0.0054062786512076855,
+ "learning_rate": 4.586574501583229e-05,
+ "loss": 3.6090235710144043,
+ "step": 3438
+ },
+ {
+ "epoch": 47.76855895196506,
+ "grad_norm": 0.005483254324644804,
+ "learning_rate": 4.5672012539937675e-05,
+ "loss": 3.647038459777832,
+ "step": 3439
+ },
+ {
+ "epoch": 47.7825327510917,
+ "grad_norm": 0.005621953401714563,
+ "learning_rate": 4.547946387506337e-05,
+ "loss": 3.64411997795105,
+ "step": 3440
+ },
+ {
+ "epoch": 47.79650655021834,
+ "grad_norm": 0.005623196717351675,
+ "learning_rate": 4.5288099176342185e-05,
+ "loss": 3.618316173553467,
+ "step": 3441
+ },
+ {
+ "epoch": 47.81048034934498,
+ "grad_norm": 0.005601753015071154,
+ "learning_rate": 4.509791859795321e-05,
+ "loss": 3.644348621368408,
+ "step": 3442
+ },
+ {
+ "epoch": 47.82445414847162,
+ "grad_norm": 0.005769635550677776,
+ "learning_rate": 4.490892229312188e-05,
+ "loss": 3.6343932151794434,
+ "step": 3443
+ },
+ {
+ "epoch": 47.838427947598255,
+ "grad_norm": 0.005564125254750252,
+ "learning_rate": 4.472111041411859e-05,
+ "loss": 3.6716389656066895,
+ "step": 3444
+ },
+ {
+ "epoch": 47.852401746724894,
+ "grad_norm": 0.0054967706091701984,
+ "learning_rate": 4.453448311226028e-05,
+ "loss": 3.6779255867004395,
+ "step": 3445
+ },
+ {
+ "epoch": 47.866375545851525,
+ "grad_norm": 0.005589650943875313,
+ "learning_rate": 4.4349040537909155e-05,
+ "loss": 3.63474702835083,
+ "step": 3446
+ },
+ {
+ "epoch": 47.880349344978164,
+ "grad_norm": 0.005526749417185783,
+ "learning_rate": 4.416478284047299e-05,
+ "loss": 3.65665864944458,
+ "step": 3447
+ },
+ {
+ "epoch": 47.8943231441048,
+ "grad_norm": 0.005574159789830446,
+ "learning_rate": 4.398171016840464e-05,
+ "loss": 3.6104025840759277,
+ "step": 3448
+ },
+ {
+ "epoch": 47.90829694323144,
+ "grad_norm": 0.00546378456056118,
+ "learning_rate": 4.379982266920288e-05,
+ "loss": 3.6659083366394043,
+ "step": 3449
+ },
+ {
+ "epoch": 47.92227074235808,
+ "grad_norm": 0.005439163185656071,
+ "learning_rate": 4.361912048941073e-05,
+ "loss": 3.6359872817993164,
+ "step": 3450
+ },
+ {
+ "epoch": 47.93624454148472,
+ "grad_norm": 0.005478638224303722,
+ "learning_rate": 4.343960377461697e-05,
+ "loss": 3.63889217376709,
+ "step": 3451
+ },
+ {
+ "epoch": 47.95021834061135,
+ "grad_norm": 0.005485034082084894,
+ "learning_rate": 4.326127266945479e-05,
+ "loss": 3.6665172576904297,
+ "step": 3452
+ },
+ {
+ "epoch": 47.96419213973799,
+ "grad_norm": 0.006088086403906345,
+ "learning_rate": 4.3084127317602475e-05,
+ "loss": 3.684257745742798,
+ "step": 3453
+ },
+ {
+ "epoch": 47.97816593886463,
+ "grad_norm": 0.005437501240521669,
+ "learning_rate": 4.290816786178288e-05,
+ "loss": 3.6222567558288574,
+ "step": 3454
+ },
+ {
+ "epoch": 47.992139737991266,
+ "grad_norm": 0.005640770774334669,
+ "learning_rate": 4.2733394443763146e-05,
+ "loss": 3.651585578918457,
+ "step": 3455
+ },
+ {
+ "epoch": 48.0,
+ "grad_norm": 0.006981590297073126,
+ "learning_rate": 4.255980720435515e-05,
+ "loss": 3.6314165592193604,
+ "step": 3456
+ },
+ {
+ "epoch": 48.0,
+ "eval_loss": 4.384180068969727,
+ "eval_runtime": 51.5098,
+ "eval_samples_per_second": 47.408,
+ "eval_steps_per_second": 1.495,
+ "step": 3456
+ },
+ {
+ "epoch": 48.01397379912664,
+ "grad_norm": 0.0055017475970089436,
+ "learning_rate": 4.238740628341503e-05,
+ "loss": 3.588742733001709,
+ "step": 3457
+ },
+ {
+ "epoch": 48.02794759825328,
+ "grad_norm": 0.005516123026609421,
+ "learning_rate": 4.22161918198432e-05,
+ "loss": 3.6423656940460205,
+ "step": 3458
+ },
+ {
+ "epoch": 48.041921397379916,
+ "grad_norm": 0.005525878164917231,
+ "learning_rate": 4.204616395158382e-05,
+ "loss": 3.6513490676879883,
+ "step": 3459
+ },
+ {
+ "epoch": 48.05589519650655,
+ "grad_norm": 0.005577936768531799,
+ "learning_rate": 4.1877322815625474e-05,
+ "loss": 3.6643600463867188,
+ "step": 3460
+ },
+ {
+ "epoch": 48.069868995633186,
+ "grad_norm": 0.005561518482863903,
+ "learning_rate": 4.170966854800052e-05,
+ "loss": 3.66424298286438,
+ "step": 3461
+ },
+ {
+ "epoch": 48.083842794759825,
+ "grad_norm": 0.005620717070996761,
+ "learning_rate": 4.154320128378459e-05,
+ "loss": 3.6455154418945312,
+ "step": 3462
+ },
+ {
+ "epoch": 48.09781659388646,
+ "grad_norm": 0.00556279718875885,
+ "learning_rate": 4.137792115709756e-05,
+ "loss": 3.634735584259033,
+ "step": 3463
+ },
+ {
+ "epoch": 48.1117903930131,
+ "grad_norm": 0.005605190526694059,
+ "learning_rate": 4.1213828301102755e-05,
+ "loss": 3.610032081604004,
+ "step": 3464
+ },
+ {
+ "epoch": 48.12576419213974,
+ "grad_norm": 0.005742431152611971,
+ "learning_rate": 4.1050922848006605e-05,
+ "loss": 3.6297264099121094,
+ "step": 3465
+ },
+ {
+ "epoch": 48.13973799126637,
+ "grad_norm": 0.005596380680799484,
+ "learning_rate": 4.088920492905929e-05,
+ "loss": 3.678337574005127,
+ "step": 3466
+ },
+ {
+ "epoch": 48.15371179039301,
+ "grad_norm": 0.0056046731770038605,
+ "learning_rate": 4.0728674674553794e-05,
+ "loss": 3.6512162685394287,
+ "step": 3467
+ },
+ {
+ "epoch": 48.16768558951965,
+ "grad_norm": 0.005494440905749798,
+ "learning_rate": 4.0569332213826826e-05,
+ "loss": 3.6533617973327637,
+ "step": 3468
+ },
+ {
+ "epoch": 48.18165938864629,
+ "grad_norm": 0.005756898783147335,
+ "learning_rate": 4.041117767525725e-05,
+ "loss": 3.649506092071533,
+ "step": 3469
+ },
+ {
+ "epoch": 48.19563318777293,
+ "grad_norm": 0.005806957837194204,
+ "learning_rate": 4.025421118626767e-05,
+ "loss": 3.695404529571533,
+ "step": 3470
+ },
+ {
+ "epoch": 48.209606986899566,
+ "grad_norm": 0.005643757060170174,
+ "learning_rate": 4.009843287332282e-05,
+ "loss": 3.656587600708008,
+ "step": 3471
+ },
+ {
+ "epoch": 48.223580786026204,
+ "grad_norm": 0.005467074923217297,
+ "learning_rate": 3.9943842861930864e-05,
+ "loss": 3.671684741973877,
+ "step": 3472
+ },
+ {
+ "epoch": 48.237554585152836,
+ "grad_norm": 0.005573616363108158,
+ "learning_rate": 3.9790441276642075e-05,
+ "loss": 3.661931037902832,
+ "step": 3473
+ },
+ {
+ "epoch": 48.251528384279474,
+ "grad_norm": 0.005787190515547991,
+ "learning_rate": 3.963822824104919e-05,
+ "loss": 3.629055976867676,
+ "step": 3474
+ },
+ {
+ "epoch": 48.26550218340611,
+ "grad_norm": 0.005627274047583342,
+ "learning_rate": 3.948720387778753e-05,
+ "loss": 3.6431307792663574,
+ "step": 3475
+ },
+ {
+ "epoch": 48.27947598253275,
+ "grad_norm": 0.005619972012937069,
+ "learning_rate": 3.933736830853473e-05,
+ "loss": 3.674844741821289,
+ "step": 3476
+ },
+ {
+ "epoch": 48.29344978165939,
+ "grad_norm": 0.0055321259424090385,
+ "learning_rate": 3.918872165401068e-05,
+ "loss": 3.6465535163879395,
+ "step": 3477
+ },
+ {
+ "epoch": 48.30742358078603,
+ "grad_norm": 0.005573419854044914,
+ "learning_rate": 3.904126403397724e-05,
+ "loss": 3.616992950439453,
+ "step": 3478
+ },
+ {
+ "epoch": 48.32139737991266,
+ "grad_norm": 0.005475590471178293,
+ "learning_rate": 3.889499556723837e-05,
+ "loss": 3.624130964279175,
+ "step": 3479
+ },
+ {
+ "epoch": 48.3353711790393,
+ "grad_norm": 0.005683086346834898,
+ "learning_rate": 3.874991637163997e-05,
+ "loss": 3.632981538772583,
+ "step": 3480
+ },
+ {
+ "epoch": 48.34934497816594,
+ "grad_norm": 0.005622202064841986,
+ "learning_rate": 3.860602656406961e-05,
+ "loss": 3.6468749046325684,
+ "step": 3481
+ },
+ {
+ "epoch": 48.36331877729258,
+ "grad_norm": 0.005536191631108522,
+ "learning_rate": 3.84633262604566e-05,
+ "loss": 3.633355140686035,
+ "step": 3482
+ },
+ {
+ "epoch": 48.377292576419215,
+ "grad_norm": 0.00547434389591217,
+ "learning_rate": 3.8321815575772385e-05,
+ "loss": 3.6079788208007812,
+ "step": 3483
+ },
+ {
+ "epoch": 48.391266375545854,
+ "grad_norm": 0.005573789123445749,
+ "learning_rate": 3.818149462402937e-05,
+ "loss": 3.6123361587524414,
+ "step": 3484
+ },
+ {
+ "epoch": 48.40524017467249,
+ "grad_norm": 0.00561481062322855,
+ "learning_rate": 3.8042363518281414e-05,
+ "loss": 3.664466142654419,
+ "step": 3485
+ },
+ {
+ "epoch": 48.419213973799124,
+ "grad_norm": 0.005487846210598946,
+ "learning_rate": 3.790442237062417e-05,
+ "loss": 3.655763626098633,
+ "step": 3486
+ },
+ {
+ "epoch": 48.43318777292576,
+ "grad_norm": 0.005416626576334238,
+ "learning_rate": 3.7767671292194405e-05,
+ "loss": 3.634082317352295,
+ "step": 3487
+ },
+ {
+ "epoch": 48.4471615720524,
+ "grad_norm": 0.005675632040947676,
+ "learning_rate": 3.76321103931697e-05,
+ "loss": 3.6449050903320312,
+ "step": 3488
+ },
+ {
+ "epoch": 48.46113537117904,
+ "grad_norm": 0.0055493442341685295,
+ "learning_rate": 3.749773978276924e-05,
+ "loss": 3.6530330181121826,
+ "step": 3489
+ },
+ {
+ "epoch": 48.47510917030568,
+ "grad_norm": 0.005522808991372585,
+ "learning_rate": 3.7364559569253025e-05,
+ "loss": 3.6732282638549805,
+ "step": 3490
+ },
+ {
+ "epoch": 48.48908296943232,
+ "grad_norm": 0.005541914142668247,
+ "learning_rate": 3.7232569859921836e-05,
+ "loss": 3.6114039421081543,
+ "step": 3491
+ },
+ {
+ "epoch": 48.50305676855895,
+ "grad_norm": 0.005627194419503212,
+ "learning_rate": 3.710177076111743e-05,
+ "loss": 3.6930434703826904,
+ "step": 3492
+ },
+ {
+ "epoch": 48.51703056768559,
+ "grad_norm": 0.005525750108063221,
+ "learning_rate": 3.697216237822202e-05,
+ "loss": 3.6784677505493164,
+ "step": 3493
+ },
+ {
+ "epoch": 48.531004366812226,
+ "grad_norm": 0.005604377947747707,
+ "learning_rate": 3.6843744815659275e-05,
+ "loss": 3.631898880004883,
+ "step": 3494
+ },
+ {
+ "epoch": 48.544978165938865,
+ "grad_norm": 0.005518812220543623,
+ "learning_rate": 3.671651817689269e-05,
+ "loss": 3.663426399230957,
+ "step": 3495
+ },
+ {
+ "epoch": 48.5589519650655,
+ "grad_norm": 0.005526282824575901,
+ "learning_rate": 3.6590482564426204e-05,
+ "loss": 3.630021572113037,
+ "step": 3496
+ },
+ {
+ "epoch": 48.57292576419214,
+ "grad_norm": 0.005505071487277746,
+ "learning_rate": 3.646563807980474e-05,
+ "loss": 3.6589057445526123,
+ "step": 3497
+ },
+ {
+ "epoch": 48.58689956331878,
+ "grad_norm": 0.00532246520742774,
+ "learning_rate": 3.6341984823613016e-05,
+ "loss": 3.640037775039673,
+ "step": 3498
+ },
+ {
+ "epoch": 48.60087336244541,
+ "grad_norm": 0.005556885153055191,
+ "learning_rate": 3.6219522895476545e-05,
+ "loss": 3.6187455654144287,
+ "step": 3499
+ },
+ {
+ "epoch": 48.61484716157205,
+ "grad_norm": 0.005715913604944944,
+ "learning_rate": 3.609825239406033e-05,
+ "loss": 3.633690118789673,
+ "step": 3500
+ },
+ {
+ "epoch": 48.62882096069869,
+ "grad_norm": 0.00542859872803092,
+ "learning_rate": 3.5978173417069984e-05,
+ "loss": 3.6094117164611816,
+ "step": 3501
+ },
+ {
+ "epoch": 48.64279475982533,
+ "grad_norm": 0.005408013239502907,
+ "learning_rate": 3.585928606125113e-05,
+ "loss": 3.680063247680664,
+ "step": 3502
+ },
+ {
+ "epoch": 48.65676855895197,
+ "grad_norm": 0.005575202405452728,
+ "learning_rate": 3.574159042238883e-05,
+ "loss": 3.6290431022644043,
+ "step": 3503
+ },
+ {
+ "epoch": 48.670742358078606,
+ "grad_norm": 0.005622019059956074,
+ "learning_rate": 3.562508659530864e-05,
+ "loss": 3.6656417846679688,
+ "step": 3504
+ },
+ {
+ "epoch": 48.68471615720524,
+ "grad_norm": 0.005383938550949097,
+ "learning_rate": 3.5509774673875424e-05,
+ "loss": 3.638788938522339,
+ "step": 3505
+ },
+ {
+ "epoch": 48.698689956331876,
+ "grad_norm": 0.005429980810731649,
+ "learning_rate": 3.539565475099401e-05,
+ "loss": 3.660780191421509,
+ "step": 3506
+ },
+ {
+ "epoch": 48.712663755458514,
+ "grad_norm": 0.005457386840134859,
+ "learning_rate": 3.528272691860856e-05,
+ "loss": 3.6496505737304688,
+ "step": 3507
+ },
+ {
+ "epoch": 48.72663755458515,
+ "grad_norm": 0.005568833090364933,
+ "learning_rate": 3.517099126770333e-05,
+ "loss": 3.657479763031006,
+ "step": 3508
+ },
+ {
+ "epoch": 48.74061135371179,
+ "grad_norm": 0.005413518752902746,
+ "learning_rate": 3.506044788830128e-05,
+ "loss": 3.681161642074585,
+ "step": 3509
+ },
+ {
+ "epoch": 48.75458515283843,
+ "grad_norm": 0.005438497755676508,
+ "learning_rate": 3.4951096869465646e-05,
+ "loss": 3.6673734188079834,
+ "step": 3510
+ },
+ {
+ "epoch": 48.76855895196506,
+ "grad_norm": 0.005298427771776915,
+ "learning_rate": 3.484293829929833e-05,
+ "loss": 3.6321797370910645,
+ "step": 3511
+ },
+ {
+ "epoch": 48.7825327510917,
+ "grad_norm": 0.0056403265334665775,
+ "learning_rate": 3.473597226494069e-05,
+ "loss": 3.625696897506714,
+ "step": 3512
+ },
+ {
+ "epoch": 48.79650655021834,
+ "grad_norm": 0.005573336500674486,
+ "learning_rate": 3.463019885257344e-05,
+ "loss": 3.656543254852295,
+ "step": 3513
+ },
+ {
+ "epoch": 48.81048034934498,
+ "grad_norm": 0.005373518913984299,
+ "learning_rate": 3.4525618147416384e-05,
+ "loss": 3.6582937240600586,
+ "step": 3514
+ },
+ {
+ "epoch": 48.82445414847162,
+ "grad_norm": 0.005456839222460985,
+ "learning_rate": 3.442223023372804e-05,
+ "loss": 3.6304612159729004,
+ "step": 3515
+ },
+ {
+ "epoch": 48.838427947598255,
+ "grad_norm": 0.0055037084966897964,
+ "learning_rate": 3.432003519480653e-05,
+ "loss": 3.6726675033569336,
+ "step": 3516
+ },
+ {
+ "epoch": 48.852401746724894,
+ "grad_norm": 0.005476827267557383,
+ "learning_rate": 3.421903311298831e-05,
+ "loss": 3.6638195514678955,
+ "step": 3517
+ },
+ {
+ "epoch": 48.866375545851525,
+ "grad_norm": 0.005663580726832151,
+ "learning_rate": 3.411922406964899e-05,
+ "loss": 3.6422252655029297,
+ "step": 3518
+ },
+ {
+ "epoch": 48.880349344978164,
+ "grad_norm": 0.005374568048864603,
+ "learning_rate": 3.402060814520299e-05,
+ "loss": 3.6490671634674072,
+ "step": 3519
+ },
+ {
+ "epoch": 48.8943231441048,
+ "grad_norm": 0.005449265707284212,
+ "learning_rate": 3.392318541910323e-05,
+ "loss": 3.6136131286621094,
+ "step": 3520
+ },
+ {
+ "epoch": 48.90829694323144,
+ "grad_norm": 0.005630240309983492,
+ "learning_rate": 3.382695596984144e-05,
+ "loss": 3.677155017852783,
+ "step": 3521
+ },
+ {
+ "epoch": 48.92227074235808,
+ "grad_norm": 0.005428317468613386,
+ "learning_rate": 3.373191987494817e-05,
+ "loss": 3.648580312728882,
+ "step": 3522
+ },
+ {
+ "epoch": 48.93624454148472,
+ "grad_norm": 0.005498172715306282,
+ "learning_rate": 3.363807721099197e-05,
+ "loss": 3.6392035484313965,
+ "step": 3523
+ },
+ {
+ "epoch": 48.95021834061135,
+ "grad_norm": 0.005650228820741177,
+ "learning_rate": 3.354542805358054e-05,
+ "loss": 3.623358964920044,
+ "step": 3524
+ },
+ {
+ "epoch": 48.96419213973799,
+ "grad_norm": 0.005477136466652155,
+ "learning_rate": 3.3453972477359395e-05,
+ "loss": 3.6367993354797363,
+ "step": 3525
+ },
+ {
+ "epoch": 48.97816593886463,
+ "grad_norm": 0.005379918497055769,
+ "learning_rate": 3.3363710556012716e-05,
+ "loss": 3.6611275672912598,
+ "step": 3526
+ },
+ {
+ "epoch": 48.992139737991266,
+ "grad_norm": 0.00544297369197011,
+ "learning_rate": 3.327464236226285e-05,
+ "loss": 3.6238505840301514,
+ "step": 3527
+ },
+ {
+ "epoch": 49.0,
+ "grad_norm": 0.00707196444272995,
+ "learning_rate": 3.3186767967870446e-05,
+ "loss": 3.664491653442383,
+ "step": 3528
+ },
+ {
+ "epoch": 49.0,
+ "eval_loss": 4.384829521179199,
+ "eval_runtime": 49.1609,
+ "eval_samples_per_second": 49.674,
+ "eval_steps_per_second": 1.566,
+ "step": 3528
+ },
+ {
+ "epoch": 49.01397379912664,
+ "grad_norm": 0.005620039068162441,
+ "learning_rate": 3.3100087443634336e-05,
+ "loss": 3.671109676361084,
+ "step": 3529
+ },
+ {
+ "epoch": 49.02794759825328,
+ "grad_norm": 0.0053699444979429245,
+ "learning_rate": 3.301460085939133e-05,
+ "loss": 3.6248927116394043,
+ "step": 3530
+ },
+ {
+ "epoch": 49.041921397379916,
+ "grad_norm": 0.005503964144736528,
+ "learning_rate": 3.2930308284016734e-05,
+ "loss": 3.653135299682617,
+ "step": 3531
+ },
+ {
+ "epoch": 49.05589519650655,
+ "grad_norm": 0.00538219278678298,
+ "learning_rate": 3.284720978542336e-05,
+ "loss": 3.6291279792785645,
+ "step": 3532
+ },
+ {
+ "epoch": 49.069868995633186,
+ "grad_norm": 0.005521378945559263,
+ "learning_rate": 3.2765305430562e-05,
+ "loss": 3.6505486965179443,
+ "step": 3533
+ },
+ {
+ "epoch": 49.083842794759825,
+ "grad_norm": 0.005552167072892189,
+ "learning_rate": 3.268459528542179e-05,
+ "loss": 3.6507344245910645,
+ "step": 3534
+ },
+ {
+ "epoch": 49.09781659388646,
+ "grad_norm": 0.005520711652934551,
+ "learning_rate": 3.260507941502933e-05,
+ "loss": 3.63080096244812,
+ "step": 3535
+ },
+ {
+ "epoch": 49.1117903930131,
+ "grad_norm": 0.005630863830447197,
+ "learning_rate": 3.25267578834491e-05,
+ "loss": 3.667436361312866,
+ "step": 3536
+ },
+ {
+ "epoch": 49.12576419213974,
+ "grad_norm": 0.005421407055109739,
+ "learning_rate": 3.244963075378337e-05,
+ "loss": 3.6779494285583496,
+ "step": 3537
+ },
+ {
+ "epoch": 49.13973799126637,
+ "grad_norm": 0.005393749102950096,
+ "learning_rate": 3.237369808817209e-05,
+ "loss": 3.6451895236968994,
+ "step": 3538
+ },
+ {
+ "epoch": 49.15371179039301,
+ "grad_norm": 0.005522069521248341,
+ "learning_rate": 3.229895994779271e-05,
+ "loss": 3.670633554458618,
+ "step": 3539
+ },
+ {
+ "epoch": 49.16768558951965,
+ "grad_norm": 0.0055035874247550964,
+ "learning_rate": 3.222541639286068e-05,
+ "loss": 3.606834888458252,
+ "step": 3540
+ },
+ {
+ "epoch": 49.18165938864629,
+ "grad_norm": 0.005393242929130793,
+ "learning_rate": 3.215306748262843e-05,
+ "loss": 3.611767292022705,
+ "step": 3541
+ },
+ {
+ "epoch": 49.19563318777293,
+ "grad_norm": 0.005510237067937851,
+ "learning_rate": 3.2081913275386245e-05,
+ "loss": 3.6726202964782715,
+ "step": 3542
+ },
+ {
+ "epoch": 49.209606986899566,
+ "grad_norm": 0.005437937099486589,
+ "learning_rate": 3.201195382846191e-05,
+ "loss": 3.676589250564575,
+ "step": 3543
+ },
+ {
+ "epoch": 49.223580786026204,
+ "grad_norm": 0.0056598675437271595,
+ "learning_rate": 3.1943189198220366e-05,
+ "loss": 3.663947105407715,
+ "step": 3544
+ },
+ {
+ "epoch": 49.237554585152836,
+ "grad_norm": 0.005362293682992458,
+ "learning_rate": 3.1875619440064064e-05,
+ "loss": 3.6216440200805664,
+ "step": 3545
+ },
+ {
+ "epoch": 49.251528384279474,
+ "grad_norm": 0.005413597449660301,
+ "learning_rate": 3.1809244608432616e-05,
+ "loss": 3.6451244354248047,
+ "step": 3546
+ },
+ {
+ "epoch": 49.26550218340611,
+ "grad_norm": 0.00545527646318078,
+ "learning_rate": 3.174406475680331e-05,
+ "loss": 3.6419572830200195,
+ "step": 3547
+ },
+ {
+ "epoch": 49.27947598253275,
+ "grad_norm": 0.005393492989242077,
+ "learning_rate": 3.16800799376901e-05,
+ "loss": 3.6303319931030273,
+ "step": 3548
+ },
+ {
+ "epoch": 49.29344978165939,
+ "grad_norm": 0.005528928246349096,
+ "learning_rate": 3.1617290202644615e-05,
+ "loss": 3.631877899169922,
+ "step": 3549
+ },
+ {
+ "epoch": 49.30742358078603,
+ "grad_norm": 0.005303470883518457,
+ "learning_rate": 3.155569560225531e-05,
+ "loss": 3.6186366081237793,
+ "step": 3550
+ },
+ {
+ "epoch": 49.32139737991266,
+ "grad_norm": 0.005345781799405813,
+ "learning_rate": 3.1495296186147835e-05,
+ "loss": 3.636327028274536,
+ "step": 3551
+ },
+ {
+ "epoch": 49.3353711790393,
+ "grad_norm": 0.00533433398231864,
+ "learning_rate": 3.143609200298481e-05,
+ "loss": 3.612666606903076,
+ "step": 3552
+ },
+ {
+ "epoch": 49.34934497816594,
+ "grad_norm": 0.005468505900353193,
+ "learning_rate": 3.137808310046605e-05,
+ "loss": 3.6335525512695312,
+ "step": 3553
+ },
+ {
+ "epoch": 49.36331877729258,
+ "grad_norm": 0.005498886574059725,
+ "learning_rate": 3.132126952532838e-05,
+ "loss": 3.64517879486084,
+ "step": 3554
+ },
+ {
+ "epoch": 49.377292576419215,
+ "grad_norm": 0.005366173572838306,
+ "learning_rate": 3.1265651323345436e-05,
+ "loss": 3.6120004653930664,
+ "step": 3555
+ },
+ {
+ "epoch": 49.391266375545854,
+ "grad_norm": 0.005579676479101181,
+ "learning_rate": 3.121122853932773e-05,
+ "loss": 3.651233196258545,
+ "step": 3556
+ },
+ {
+ "epoch": 49.40524017467249,
+ "grad_norm": 0.005380513612180948,
+ "learning_rate": 3.115800121712275e-05,
+ "loss": 3.6802916526794434,
+ "step": 3557
+ },
+ {
+ "epoch": 49.419213973799124,
+ "grad_norm": 0.005381321534514427,
+ "learning_rate": 3.110596939961485e-05,
+ "loss": 3.606739044189453,
+ "step": 3558
+ },
+ {
+ "epoch": 49.43318777292576,
+ "grad_norm": 0.005590811837464571,
+ "learning_rate": 3.105513312872503e-05,
+ "loss": 3.6593852043151855,
+ "step": 3559
+ },
+ {
+ "epoch": 49.4471615720524,
+ "grad_norm": 0.005552604794502258,
+ "learning_rate": 3.100549244541116e-05,
+ "loss": 3.6589818000793457,
+ "step": 3560
+ },
+ {
+ "epoch": 49.46113537117904,
+ "grad_norm": 0.005272792186588049,
+ "learning_rate": 3.0957047389667774e-05,
+ "loss": 3.6500585079193115,
+ "step": 3561
+ },
+ {
+ "epoch": 49.47510917030568,
+ "grad_norm": 0.00557938776910305,
+ "learning_rate": 3.09097980005264e-05,
+ "loss": 3.6423044204711914,
+ "step": 3562
+ },
+ {
+ "epoch": 49.48908296943232,
+ "grad_norm": 0.005462596192955971,
+ "learning_rate": 3.0863744316054773e-05,
+ "loss": 3.6106159687042236,
+ "step": 3563
+ },
+ {
+ "epoch": 49.50305676855895,
+ "grad_norm": 0.005449238698929548,
+ "learning_rate": 3.081888637335776e-05,
+ "loss": 3.6468400955200195,
+ "step": 3564
+ },
+ {
+ "epoch": 49.51703056768559,
+ "grad_norm": 0.005447402596473694,
+ "learning_rate": 3.077522420857645e-05,
+ "loss": 3.624122142791748,
+ "step": 3565
+ },
+ {
+ "epoch": 49.531004366812226,
+ "grad_norm": 0.005388482473790646,
+ "learning_rate": 3.0732757856888576e-05,
+ "loss": 3.6392014026641846,
+ "step": 3566
+ },
+ {
+ "epoch": 49.544978165938865,
+ "grad_norm": 0.00541285565122962,
+ "learning_rate": 3.069148735250873e-05,
+ "loss": 3.654369354248047,
+ "step": 3567
+ },
+ {
+ "epoch": 49.5589519650655,
+ "grad_norm": 0.005579780787229538,
+ "learning_rate": 3.065141272868768e-05,
+ "loss": 3.6372079849243164,
+ "step": 3568
+ },
+ {
+ "epoch": 49.57292576419214,
+ "grad_norm": 0.005476080346852541,
+ "learning_rate": 3.061253401771287e-05,
+ "loss": 3.621183395385742,
+ "step": 3569
+ },
+ {
+ "epoch": 49.58689956331878,
+ "grad_norm": 0.005435108672827482,
+ "learning_rate": 3.057485125090808e-05,
+ "loss": 3.6125760078430176,
+ "step": 3570
+ },
+ {
+ "epoch": 49.60087336244541,
+ "grad_norm": 0.005412368569523096,
+ "learning_rate": 3.0538364458633627e-05,
+ "loss": 3.6276559829711914,
+ "step": 3571
+ },
+ {
+ "epoch": 49.61484716157205,
+ "grad_norm": 0.005578193813562393,
+ "learning_rate": 3.050307367028646e-05,
+ "loss": 3.672255516052246,
+ "step": 3572
+ },
+ {
+ "epoch": 49.62882096069869,
+ "grad_norm": 0.005397310480475426,
+ "learning_rate": 3.046897891429959e-05,
+ "loss": 3.6349916458129883,
+ "step": 3573
+ },
+ {
+ "epoch": 49.64279475982533,
+ "grad_norm": 0.005437100306153297,
+ "learning_rate": 3.0436080218142514e-05,
+ "loss": 3.641871929168701,
+ "step": 3574
+ },
+ {
+ "epoch": 49.65676855895197,
+ "grad_norm": 0.00533803878352046,
+ "learning_rate": 3.0404377608321075e-05,
+ "loss": 3.6466495990753174,
+ "step": 3575
+ },
+ {
+ "epoch": 49.670742358078606,
+ "grad_norm": 0.005386525299400091,
+ "learning_rate": 3.0373871110377803e-05,
+ "loss": 3.6554203033447266,
+ "step": 3576
+ },
+ {
+ "epoch": 49.68471615720524,
+ "grad_norm": 0.005472255405038595,
+ "learning_rate": 3.0344560748890904e-05,
+ "loss": 3.6272778511047363,
+ "step": 3577
+ },
+ {
+ "epoch": 49.698689956331876,
+ "grad_norm": 0.005554827395826578,
+ "learning_rate": 3.0316446547475262e-05,
+ "loss": 3.6589441299438477,
+ "step": 3578
+ },
+ {
+ "epoch": 49.712663755458514,
+ "grad_norm": 0.005507173947989941,
+ "learning_rate": 3.0289528528782108e-05,
+ "loss": 3.660010814666748,
+ "step": 3579
+ },
+ {
+ "epoch": 49.72663755458515,
+ "grad_norm": 0.005387808196246624,
+ "learning_rate": 3.0263806714498688e-05,
+ "loss": 3.6784188747406006,
+ "step": 3580
+ },
+ {
+ "epoch": 49.74061135371179,
+ "grad_norm": 0.005291743203997612,
+ "learning_rate": 3.0239281125348596e-05,
+ "loss": 3.665433168411255,
+ "step": 3581
+ },
+ {
+ "epoch": 49.75458515283843,
+ "grad_norm": 0.005448203068226576,
+ "learning_rate": 3.0215951781091777e-05,
+ "loss": 3.6186153888702393,
+ "step": 3582
+ },
+ {
+ "epoch": 49.76855895196506,
+ "grad_norm": 0.005570375360548496,
+ "learning_rate": 3.0193818700524177e-05,
+ "loss": 3.6409573554992676,
+ "step": 3583
+ },
+ {
+ "epoch": 49.7825327510917,
+ "grad_norm": 0.0053171273320913315,
+ "learning_rate": 3.0172881901477933e-05,
+ "loss": 3.627653121948242,
+ "step": 3584
+ },
+ {
+ "epoch": 49.79650655021834,
+ "grad_norm": 0.0054213316179811954,
+ "learning_rate": 3.0153141400821696e-05,
+ "loss": 3.6836981773376465,
+ "step": 3585
+ },
+ {
+ "epoch": 49.81048034934498,
+ "grad_norm": 0.005401757545769215,
+ "learning_rate": 3.0134597214459794e-05,
+ "loss": 3.665957450866699,
+ "step": 3586
+ },
+ {
+ "epoch": 49.82445414847162,
+ "grad_norm": 0.005455177277326584,
+ "learning_rate": 3.0117249357333078e-05,
+ "loss": 3.613741874694824,
+ "step": 3587
+ },
+ {
+ "epoch": 49.838427947598255,
+ "grad_norm": 0.005288220476359129,
+ "learning_rate": 3.0101097843418394e-05,
+ "loss": 3.66194486618042,
+ "step": 3588
+ },
+ {
+ "epoch": 49.852401746724894,
+ "grad_norm": 0.005295758601278067,
+ "learning_rate": 3.0086142685728467e-05,
+ "loss": 3.6889102458953857,
+ "step": 3589
+ },
+ {
+ "epoch": 49.866375545851525,
+ "grad_norm": 0.005356399342417717,
+ "learning_rate": 3.0072383896312843e-05,
+ "loss": 3.6496775150299072,
+ "step": 3590
+ },
+ {
+ "epoch": 49.880349344978164,
+ "grad_norm": 0.0054651242680847645,
+ "learning_rate": 3.0059821486256425e-05,
+ "loss": 3.6584572792053223,
+ "step": 3591
+ },
+ {
+ "epoch": 49.8943231441048,
+ "grad_norm": 0.005395952612161636,
+ "learning_rate": 3.004845546568064e-05,
+ "loss": 3.633230686187744,
+ "step": 3592
+ },
+ {
+ "epoch": 49.90829694323144,
+ "grad_norm": 0.005331660155206919,
+ "learning_rate": 3.003828584374276e-05,
+ "loss": 3.628465175628662,
+ "step": 3593
+ },
+ {
+ "epoch": 49.92227074235808,
+ "grad_norm": 0.0054406700655817986,
+ "learning_rate": 3.0029312628636385e-05,
+ "loss": 3.665616512298584,
+ "step": 3594
+ },
+ {
+ "epoch": 49.93624454148472,
+ "grad_norm": 0.005537047050893307,
+ "learning_rate": 3.0021535827590996e-05,
+ "loss": 3.650026321411133,
+ "step": 3595
+ },
+ {
+ "epoch": 49.95021834061135,
+ "grad_norm": 0.005354189779609442,
+ "learning_rate": 3.0014955446872235e-05,
+ "loss": 3.6486058235168457,
+ "step": 3596
+ },
+ {
+ "epoch": 49.96419213973799,
+ "grad_norm": 0.0054948581382632256,
+ "learning_rate": 3.000957149178194e-05,
+ "loss": 3.632627010345459,
+ "step": 3597
+ },
+ {
+ "epoch": 49.97816593886463,
+ "grad_norm": 0.00550198694691062,
+ "learning_rate": 3.000538396665778e-05,
+ "loss": 3.6374807357788086,
+ "step": 3598
+ },
+ {
+ "epoch": 49.992139737991266,
+ "grad_norm": 0.005545698571950197,
+ "learning_rate": 3.000239287487345e-05,
+ "loss": 3.6445484161376953,
+ "step": 3599
+ },
+ {
+ "epoch": 50.0,
+ "grad_norm": 0.006927069742232561,
+ "learning_rate": 3.0000598218838843e-05,
+ "loss": 3.6755237579345703,
+ "step": 3600
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": true
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 1.534806680666112e+18,
+ "train_batch_size": 16,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/baseline/checkpoint-3600/training_args.bin b/runs/baseline/checkpoint-3600/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7
--- /dev/null
+++ b/runs/baseline/checkpoint-3600/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97
+size 5265
diff --git a/runs/baseline/checkpoint-72/chat_template.jinja b/runs/baseline/checkpoint-72/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/baseline/checkpoint-72/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/baseline/checkpoint-72/config.json b/runs/baseline/checkpoint-72/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/baseline/checkpoint-72/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/baseline/checkpoint-72/generation_config.json b/runs/baseline/checkpoint-72/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/baseline/checkpoint-72/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/baseline/checkpoint-72/model.safetensors b/runs/baseline/checkpoint-72/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..9c5da32c04bbbd011788bcc0535c61e1d4fb4994
--- /dev/null
+++ b/runs/baseline/checkpoint-72/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b68864cea94a3e73fe9caa2a0dc90451588ffa255ab33997e5376c85df693ab2
+size 583356232
diff --git a/runs/baseline/checkpoint-72/optimizer.pt b/runs/baseline/checkpoint-72/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..0d44ac376410c76eddf7dab317ac0077f2594c35
--- /dev/null
+++ b/runs/baseline/checkpoint-72/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a5f8e15426c5c1eb16331c156404fa8cd028829039d2caf56daa8739d001fe8c
+size 1166825803
diff --git a/runs/baseline/checkpoint-72/rng_state_0.pth b/runs/baseline/checkpoint-72/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..3fc68e18ddaf65dfbdec55893d9a925ff5e43a18
--- /dev/null
+++ b/runs/baseline/checkpoint-72/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4dd7671ce88d469c49c0530724ac76b2306574002d1ecd1ca9294e41621fd96a
+size 14917
diff --git a/runs/baseline/checkpoint-72/rng_state_1.pth b/runs/baseline/checkpoint-72/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..79d9de29ae34b3b0c10ea8ba0348aeafb0c12226
--- /dev/null
+++ b/runs/baseline/checkpoint-72/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3246ef1170ccca541a03b89ad6f20e01c51eb6834a2c2211c78c71c70f896879
+size 14917
diff --git a/runs/baseline/checkpoint-72/scheduler.pt b/runs/baseline/checkpoint-72/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..d33c2d0d0a3862ff1e2785405371ddce28072435
--- /dev/null
+++ b/runs/baseline/checkpoint-72/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ae8ce601b5ae4abd83f2c331115c5c2500c7c5c7f6bc8590a847682f194a23a4
+size 1465
diff --git a/runs/baseline/checkpoint-72/tokenizer.json b/runs/baseline/checkpoint-72/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/baseline/checkpoint-72/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/baseline/checkpoint-72/tokenizer_config.json b/runs/baseline/checkpoint-72/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/baseline/checkpoint-72/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/baseline/checkpoint-72/trainer_state.json b/runs/baseline/checkpoint-72/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..7afea72c7a477c337cded595f728ef51ea45e701
--- /dev/null
+++ b/runs/baseline/checkpoint-72/trainer_state.json
@@ -0,0 +1,538 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 1.0,
+ "eval_steps": 500,
+ "global_step": 72,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.12165174633264542,
+ "learning_rate": 0.0,
+ "loss": 12.027063369750977,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.12109383195638657,
+ "learning_rate": 3e-05,
+ "loss": 12.028635025024414,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.12690064311027527,
+ "learning_rate": 6e-05,
+ "loss": 11.894524574279785,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.1537286937236786,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.672262191772461,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.11778244376182556,
+ "learning_rate": 0.00012,
+ "loss": 11.433671951293945,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.10874935984611511,
+ "learning_rate": 0.00015000000000000001,
+ "loss": 11.270309448242188,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.10641921311616898,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 11.155481338500977,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.1048506498336792,
+ "learning_rate": 0.00021000000000000004,
+ "loss": 11.066675186157227,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.10546834021806717,
+ "learning_rate": 0.00024,
+ "loss": 10.952295303344727,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.10614268481731415,
+ "learning_rate": 0.00027,
+ "loss": 10.811386108398438,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10529010742902756,
+ "learning_rate": 0.00030000000000000003,
+ "loss": 10.657712936401367,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.1051444336771965,
+ "learning_rate": 0.00033,
+ "loss": 10.47591781616211,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10420846194028854,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 10.280141830444336,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10323625057935715,
+ "learning_rate": 0.00039000000000000005,
+ "loss": 10.076366424560547,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10255401581525803,
+ "learning_rate": 0.00042000000000000007,
+ "loss": 9.856939315795898,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10151027888059616,
+ "learning_rate": 0.00045,
+ "loss": 9.632134437561035,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.1006326898932457,
+ "learning_rate": 0.00048,
+ "loss": 9.401637077331543,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.09820562601089478,
+ "learning_rate": 0.00051,
+ "loss": 9.177400588989258,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.0956626906991005,
+ "learning_rate": 0.00054,
+ "loss": 8.95518684387207,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.0922091156244278,
+ "learning_rate": 0.00057,
+ "loss": 8.75066089630127,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.08593932539224625,
+ "learning_rate": 0.0006000000000000001,
+ "loss": 8.557976722717285,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.08014766871929169,
+ "learning_rate": 0.00063,
+ "loss": 8.360604286193848,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.07381244748830795,
+ "learning_rate": 0.00066,
+ "loss": 8.210375785827637,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.06455052644014359,
+ "learning_rate": 0.0006900000000000001,
+ "loss": 8.07168960571289,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.054115474224090576,
+ "learning_rate": 0.0007199999999999999,
+ "loss": 7.960069179534912,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.043575264513492584,
+ "learning_rate": 0.00075,
+ "loss": 7.850685119628906,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.02976679615676403,
+ "learning_rate": 0.0007800000000000001,
+ "loss": 7.824377059936523,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.020552165806293488,
+ "learning_rate": 0.0008100000000000001,
+ "loss": 7.7695417404174805,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.018530743196606636,
+ "learning_rate": 0.0008400000000000001,
+ "loss": 7.782247543334961,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.02716820128262043,
+ "learning_rate": 0.00087,
+ "loss": 7.754945278167725,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.03638409823179245,
+ "learning_rate": 0.0009,
+ "loss": 7.799381256103516,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.041579652577638626,
+ "learning_rate": 0.00093,
+ "loss": 7.776629447937012,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.0304960235953331,
+ "learning_rate": 0.00096,
+ "loss": 7.749439239501953,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.053728822618722916,
+ "learning_rate": 0.00099,
+ "loss": 7.745059490203857,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.036322493106126785,
+ "learning_rate": 0.00102,
+ "loss": 7.752074718475342,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.03232959285378456,
+ "learning_rate": 0.00105,
+ "loss": 7.704257011413574,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.03246184438467026,
+ "learning_rate": 0.00108,
+ "loss": 7.667045593261719,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.036655545234680176,
+ "learning_rate": 0.00111,
+ "loss": 7.623445987701416,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.021042725071310997,
+ "learning_rate": 0.00114,
+ "loss": 7.582821846008301,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.022982975468039513,
+ "learning_rate": 0.00117,
+ "loss": 7.58036994934082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.024781864136457443,
+ "learning_rate": 0.0012000000000000001,
+ "loss": 7.5903000831604,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.03667089343070984,
+ "learning_rate": 0.00123,
+ "loss": 7.588233947753906,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.04083017632365227,
+ "learning_rate": 0.00126,
+ "loss": 7.548343658447266,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.02585385926067829,
+ "learning_rate": 0.00129,
+ "loss": 7.604833602905273,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.018793931230902672,
+ "learning_rate": 0.00132,
+ "loss": 7.56705379486084,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.025538936257362366,
+ "learning_rate": 0.00135,
+ "loss": 7.562107086181641,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.02819318138062954,
+ "learning_rate": 0.0013800000000000002,
+ "loss": 7.517213821411133,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.016239255666732788,
+ "learning_rate": 0.00141,
+ "loss": 7.4780755043029785,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.024470701813697815,
+ "learning_rate": 0.0014399999999999999,
+ "loss": 7.505887985229492,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.015499847941100597,
+ "learning_rate": 0.00147,
+ "loss": 7.484469890594482,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.017785200849175453,
+ "learning_rate": 0.0015,
+ "loss": 7.481822967529297,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.015952805057168007,
+ "learning_rate": 0.0015300000000000001,
+ "loss": 7.484804153442383,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.015771260485053062,
+ "learning_rate": 0.0015600000000000002,
+ "loss": 7.4495439529418945,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.024370530620217323,
+ "learning_rate": 0.00159,
+ "loss": 7.471208572387695,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.018173737451434135,
+ "learning_rate": 0.0016200000000000001,
+ "loss": 7.438447952270508,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.020564094185829163,
+ "learning_rate": 0.0016500000000000002,
+ "loss": 7.437922954559326,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.016154944896697998,
+ "learning_rate": 0.0016800000000000003,
+ "loss": 7.429378032684326,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.016817007213830948,
+ "learning_rate": 0.00171,
+ "loss": 7.406428337097168,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.03460947796702385,
+ "learning_rate": 0.00174,
+ "loss": 7.393331527709961,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.15410488843917847,
+ "learning_rate": 0.0017699999999999999,
+ "loss": 7.645856857299805,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.07443148642778397,
+ "learning_rate": 0.0018,
+ "loss": 7.549508094787598,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.0507572740316391,
+ "learning_rate": 0.00183,
+ "loss": 7.517921447753906,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.09682746231555939,
+ "learning_rate": 0.00186,
+ "loss": 7.543978691101074,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.2571517527103424,
+ "learning_rate": 0.00189,
+ "loss": 7.768949508666992,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.0317784883081913,
+ "learning_rate": 0.00192,
+ "loss": 7.509063720703125,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.0353970006108284,
+ "learning_rate": 0.0019500000000000001,
+ "loss": 7.522322654724121,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.03663994371891022,
+ "learning_rate": 0.00198,
+ "loss": 7.504723072052002,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.025481393560767174,
+ "learning_rate": 0.00201,
+ "loss": 7.452597141265869,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.03099915198981762,
+ "learning_rate": 0.00204,
+ "loss": 7.455441474914551,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.022921698167920113,
+ "learning_rate": 0.00207,
+ "loss": 7.4454827308654785,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.032388731837272644,
+ "learning_rate": 0.0021,
+ "loss": 7.467103958129883,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.022428609430789948,
+ "learning_rate": 0.00213,
+ "loss": 7.467540264129639,
+ "step": 72
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.069613361332224e+16,
+ "train_batch_size": 16,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/baseline/checkpoint-72/training_args.bin b/runs/baseline/checkpoint-72/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7
--- /dev/null
+++ b/runs/baseline/checkpoint-72/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97
+size 5265
diff --git a/runs/baseline/checkpoint-720/chat_template.jinja b/runs/baseline/checkpoint-720/chat_template.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230
--- /dev/null
+++ b/runs/baseline/checkpoint-720/chat_template.jinja
@@ -0,0 +1,85 @@
+{%- if tools %}
+ {{- '<|im_start|>system\n' }}
+ {%- if messages[0].role == 'system' %}
+ {{- messages[0].content + '\n\n' }}
+ {%- endif %}
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }}
+ {%- for tool in tools %}
+ {{- "\n" }}
+ {{- tool | tojson }}
+ {%- endfor %}
+ {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }}
+{%- else %}
+ {%- if messages[0].role == 'system' %}
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
+ {%- endif %}
+{%- endif %}
+{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
+{%- for message in messages[::-1] %}
+ {%- set index = (messages|length - 1) - loop.index0 %}
+ {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %}
+ {%- set ns.multi_step_tool = false %}
+ {%- set ns.last_query_index = index %}
+ {%- endif %}
+{%- endfor %}
+{%- for message in messages %}
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }}
+ {%- elif message.role == "assistant" %}
+ {%- set content = message.content %}
+ {%- set reasoning_content = '' %}
+ {%- if message.reasoning_content is defined and message.reasoning_content is not none %}
+ {%- set reasoning_content = message.reasoning_content %}
+ {%- else %}
+ {%- if '' in message.content %}
+ {%- set content = message.content.split('')[-1].lstrip('\n') %}
+ {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %}
+ {%- endif %}
+ {%- endif %}
+ {%- if loop.index0 > ns.last_query_index %}
+ {%- if loop.last or (not loop.last and reasoning_content) %}
+ {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- else %}
+ {{- '<|im_start|>' + message.role + '\n' + content }}
+ {%- endif %}
+ {%- if message.tool_calls %}
+ {%- for tool_call in message.tool_calls %}
+ {%- if (loop.first and content) or (not loop.first) %}
+ {{- '\n' }}
+ {%- endif %}
+ {%- if tool_call.function %}
+ {%- set tool_call = tool_call.function %}
+ {%- endif %}
+ {{- '\n{"name": "' }}
+ {{- tool_call.name }}
+ {{- '", "arguments": ' }}
+ {%- if tool_call.arguments is string %}
+ {{- tool_call.arguments }}
+ {%- else %}
+ {{- tool_call.arguments | tojson }}
+ {%- endif %}
+ {{- '}\n' }}
+ {%- endfor %}
+ {%- endif %}
+ {{- '<|im_end|>\n' }}
+ {%- elif message.role == "tool" %}
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
+ {{- '<|im_start|>user' }}
+ {%- endif %}
+ {{- '\n\n' }}
+ {{- message.content }}
+ {{- '\n' }}
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
+ {{- '<|im_end|>\n' }}
+ {%- endif %}
+ {%- endif %}
+{%- endfor %}
+{%- if add_generation_prompt %}
+ {{- '<|im_start|>assistant\n' }}
+ {%- if enable_thinking is defined and enable_thinking is false %}
+ {{- '\n\n\n\n' }}
+ {%- endif %}
+{%- endif %}
\ No newline at end of file
diff --git a/runs/baseline/checkpoint-720/config.json b/runs/baseline/checkpoint-720/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de
--- /dev/null
+++ b/runs/baseline/checkpoint-720/config.json
@@ -0,0 +1,32 @@
+{
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": null,
+ "dtype": "float32",
+ "eos_token_id": 151645,
+ "head_dim": 128,
+ "hidden_act": "silu",
+ "hidden_size": 512,
+ "initializer_range": 0.02,
+ "intermediate_size": 1536,
+ "max_position_embeddings": 2048,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 4,
+ "num_hidden_layers": 20,
+ "num_key_value_heads": 4,
+ "pad_token_id": 151645,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-06,
+ "rope_parameters": {
+ "rope_theta": 10000.0,
+ "rope_type": "default"
+ },
+ "tie_word_embeddings": true,
+ "transformers_version": "5.5.0",
+ "use_cache": false,
+ "vocab_size": 151671
+}
diff --git a/runs/baseline/checkpoint-720/generation_config.json b/runs/baseline/checkpoint-720/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3
--- /dev/null
+++ b/runs/baseline/checkpoint-720/generation_config.json
@@ -0,0 +1,11 @@
+{
+ "_from_model_config": true,
+ "eos_token_id": [
+ 151645
+ ],
+ "output_attentions": false,
+ "output_hidden_states": false,
+ "pad_token_id": 151645,
+ "transformers_version": "5.5.0",
+ "use_cache": true
+}
diff --git a/runs/baseline/checkpoint-720/model.safetensors b/runs/baseline/checkpoint-720/model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..e76e4b365029a6aa0f0daa39af6bb2ac736fbb1c
--- /dev/null
+++ b/runs/baseline/checkpoint-720/model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:91f481106cb2d8dcc4f7d57ef7ae377bb737fa8611a827eec3ed1e957fad3d2c
+size 583356232
diff --git a/runs/baseline/checkpoint-720/optimizer.pt b/runs/baseline/checkpoint-720/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..b8c25eac628166256f178d45c0366ae78755dc5a
--- /dev/null
+++ b/runs/baseline/checkpoint-720/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e60c4d7df3ee39416f5e537e1935b44bc5d1088ee3a74cdabef6b771df36e9d7
+size 1166825803
diff --git a/runs/baseline/checkpoint-720/rng_state_0.pth b/runs/baseline/checkpoint-720/rng_state_0.pth
new file mode 100644
index 0000000000000000000000000000000000000000..64991ceae73e5d3389476a9b448c3cda96f73b10
--- /dev/null
+++ b/runs/baseline/checkpoint-720/rng_state_0.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ed1b9484e9cec07c14c677aaf97ebf9dff6fa34fc12bae83c65f98c9d2c9ad0d
+size 14917
diff --git a/runs/baseline/checkpoint-720/rng_state_1.pth b/runs/baseline/checkpoint-720/rng_state_1.pth
new file mode 100644
index 0000000000000000000000000000000000000000..fee3b6b64b847dcce05678311e2696b568d3185a
--- /dev/null
+++ b/runs/baseline/checkpoint-720/rng_state_1.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d30790dc4fc4171a789605cefdf649fd85c4d4afb59efd35ee6fea409dad4915
+size 14917
diff --git a/runs/baseline/checkpoint-720/scheduler.pt b/runs/baseline/checkpoint-720/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..454d8217902b1d6b2ad959de62cd253bdcce51c8
--- /dev/null
+++ b/runs/baseline/checkpoint-720/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:29fc245c65149711c0ae7f26d9977031181d64eae033b8ce0b653bc6512fd5e0
+size 1465
diff --git a/runs/baseline/checkpoint-720/tokenizer.json b/runs/baseline/checkpoint-720/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15
--- /dev/null
+++ b/runs/baseline/checkpoint-720/tokenizer.json
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb
+size 11424079
diff --git a/runs/baseline/checkpoint-720/tokenizer_config.json b/runs/baseline/checkpoint-720/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150
--- /dev/null
+++ b/runs/baseline/checkpoint-720/tokenizer_config.json
@@ -0,0 +1,18 @@
+{
+ "add_prefix_space": false,
+ "backend": "tokenizers",
+ "bos_token": null,
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "<|im_end|>",
+ "errors": "replace",
+ "extra_special_tokens": [
+ "<|l2r_pred|>",
+ "<|r2l_pred|>"
+ ],
+ "is_local": false,
+ "model_max_length": 131072,
+ "pad_token": "<|im_end|>",
+ "split_special_tokens": false,
+ "tokenizer_class": "Qwen2Tokenizer",
+ "unk_token": null
+}
diff --git a/runs/baseline/checkpoint-720/trainer_state.json b/runs/baseline/checkpoint-720/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..d7ceb5c73632a2a8723625795bfb0dfe919c03d1
--- /dev/null
+++ b/runs/baseline/checkpoint-720/trainer_state.json
@@ -0,0 +1,5146 @@
+{
+ "best_global_step": null,
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 10.0,
+ "eval_steps": 500,
+ "global_step": 720,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.013973799126637555,
+ "grad_norm": 0.12165174633264542,
+ "learning_rate": 0.0,
+ "loss": 12.027063369750977,
+ "step": 1
+ },
+ {
+ "epoch": 0.02794759825327511,
+ "grad_norm": 0.12109383195638657,
+ "learning_rate": 3e-05,
+ "loss": 12.028635025024414,
+ "step": 2
+ },
+ {
+ "epoch": 0.04192139737991266,
+ "grad_norm": 0.12690064311027527,
+ "learning_rate": 6e-05,
+ "loss": 11.894524574279785,
+ "step": 3
+ },
+ {
+ "epoch": 0.05589519650655022,
+ "grad_norm": 0.1537286937236786,
+ "learning_rate": 8.999999999999999e-05,
+ "loss": 11.672262191772461,
+ "step": 4
+ },
+ {
+ "epoch": 0.06986899563318777,
+ "grad_norm": 0.11778244376182556,
+ "learning_rate": 0.00012,
+ "loss": 11.433671951293945,
+ "step": 5
+ },
+ {
+ "epoch": 0.08384279475982533,
+ "grad_norm": 0.10874935984611511,
+ "learning_rate": 0.00015000000000000001,
+ "loss": 11.270309448242188,
+ "step": 6
+ },
+ {
+ "epoch": 0.09781659388646288,
+ "grad_norm": 0.10641921311616898,
+ "learning_rate": 0.00017999999999999998,
+ "loss": 11.155481338500977,
+ "step": 7
+ },
+ {
+ "epoch": 0.11179039301310044,
+ "grad_norm": 0.1048506498336792,
+ "learning_rate": 0.00021000000000000004,
+ "loss": 11.066675186157227,
+ "step": 8
+ },
+ {
+ "epoch": 0.125764192139738,
+ "grad_norm": 0.10546834021806717,
+ "learning_rate": 0.00024,
+ "loss": 10.952295303344727,
+ "step": 9
+ },
+ {
+ "epoch": 0.13973799126637554,
+ "grad_norm": 0.10614268481731415,
+ "learning_rate": 0.00027,
+ "loss": 10.811386108398438,
+ "step": 10
+ },
+ {
+ "epoch": 0.1537117903930131,
+ "grad_norm": 0.10529010742902756,
+ "learning_rate": 0.00030000000000000003,
+ "loss": 10.657712936401367,
+ "step": 11
+ },
+ {
+ "epoch": 0.16768558951965065,
+ "grad_norm": 0.1051444336771965,
+ "learning_rate": 0.00033,
+ "loss": 10.47591781616211,
+ "step": 12
+ },
+ {
+ "epoch": 0.18165938864628822,
+ "grad_norm": 0.10420846194028854,
+ "learning_rate": 0.00035999999999999997,
+ "loss": 10.280141830444336,
+ "step": 13
+ },
+ {
+ "epoch": 0.19563318777292577,
+ "grad_norm": 0.10323625057935715,
+ "learning_rate": 0.00039000000000000005,
+ "loss": 10.076366424560547,
+ "step": 14
+ },
+ {
+ "epoch": 0.2096069868995633,
+ "grad_norm": 0.10255401581525803,
+ "learning_rate": 0.00042000000000000007,
+ "loss": 9.856939315795898,
+ "step": 15
+ },
+ {
+ "epoch": 0.22358078602620088,
+ "grad_norm": 0.10151027888059616,
+ "learning_rate": 0.00045,
+ "loss": 9.632134437561035,
+ "step": 16
+ },
+ {
+ "epoch": 0.23755458515283842,
+ "grad_norm": 0.1006326898932457,
+ "learning_rate": 0.00048,
+ "loss": 9.401637077331543,
+ "step": 17
+ },
+ {
+ "epoch": 0.251528384279476,
+ "grad_norm": 0.09820562601089478,
+ "learning_rate": 0.00051,
+ "loss": 9.177400588989258,
+ "step": 18
+ },
+ {
+ "epoch": 0.26550218340611353,
+ "grad_norm": 0.0956626906991005,
+ "learning_rate": 0.00054,
+ "loss": 8.95518684387207,
+ "step": 19
+ },
+ {
+ "epoch": 0.2794759825327511,
+ "grad_norm": 0.0922091156244278,
+ "learning_rate": 0.00057,
+ "loss": 8.75066089630127,
+ "step": 20
+ },
+ {
+ "epoch": 0.2934497816593886,
+ "grad_norm": 0.08593932539224625,
+ "learning_rate": 0.0006000000000000001,
+ "loss": 8.557976722717285,
+ "step": 21
+ },
+ {
+ "epoch": 0.3074235807860262,
+ "grad_norm": 0.08014766871929169,
+ "learning_rate": 0.00063,
+ "loss": 8.360604286193848,
+ "step": 22
+ },
+ {
+ "epoch": 0.32139737991266376,
+ "grad_norm": 0.07381244748830795,
+ "learning_rate": 0.00066,
+ "loss": 8.210375785827637,
+ "step": 23
+ },
+ {
+ "epoch": 0.3353711790393013,
+ "grad_norm": 0.06455052644014359,
+ "learning_rate": 0.0006900000000000001,
+ "loss": 8.07168960571289,
+ "step": 24
+ },
+ {
+ "epoch": 0.34934497816593885,
+ "grad_norm": 0.054115474224090576,
+ "learning_rate": 0.0007199999999999999,
+ "loss": 7.960069179534912,
+ "step": 25
+ },
+ {
+ "epoch": 0.36331877729257644,
+ "grad_norm": 0.043575264513492584,
+ "learning_rate": 0.00075,
+ "loss": 7.850685119628906,
+ "step": 26
+ },
+ {
+ "epoch": 0.377292576419214,
+ "grad_norm": 0.02976679615676403,
+ "learning_rate": 0.0007800000000000001,
+ "loss": 7.824377059936523,
+ "step": 27
+ },
+ {
+ "epoch": 0.39126637554585153,
+ "grad_norm": 0.020552165806293488,
+ "learning_rate": 0.0008100000000000001,
+ "loss": 7.7695417404174805,
+ "step": 28
+ },
+ {
+ "epoch": 0.4052401746724891,
+ "grad_norm": 0.018530743196606636,
+ "learning_rate": 0.0008400000000000001,
+ "loss": 7.782247543334961,
+ "step": 29
+ },
+ {
+ "epoch": 0.4192139737991266,
+ "grad_norm": 0.02716820128262043,
+ "learning_rate": 0.00087,
+ "loss": 7.754945278167725,
+ "step": 30
+ },
+ {
+ "epoch": 0.4331877729257642,
+ "grad_norm": 0.03638409823179245,
+ "learning_rate": 0.0009,
+ "loss": 7.799381256103516,
+ "step": 31
+ },
+ {
+ "epoch": 0.44716157205240176,
+ "grad_norm": 0.041579652577638626,
+ "learning_rate": 0.00093,
+ "loss": 7.776629447937012,
+ "step": 32
+ },
+ {
+ "epoch": 0.4611353711790393,
+ "grad_norm": 0.0304960235953331,
+ "learning_rate": 0.00096,
+ "loss": 7.749439239501953,
+ "step": 33
+ },
+ {
+ "epoch": 0.47510917030567684,
+ "grad_norm": 0.053728822618722916,
+ "learning_rate": 0.00099,
+ "loss": 7.745059490203857,
+ "step": 34
+ },
+ {
+ "epoch": 0.4890829694323144,
+ "grad_norm": 0.036322493106126785,
+ "learning_rate": 0.00102,
+ "loss": 7.752074718475342,
+ "step": 35
+ },
+ {
+ "epoch": 0.503056768558952,
+ "grad_norm": 0.03232959285378456,
+ "learning_rate": 0.00105,
+ "loss": 7.704257011413574,
+ "step": 36
+ },
+ {
+ "epoch": 0.5170305676855895,
+ "grad_norm": 0.03246184438467026,
+ "learning_rate": 0.00108,
+ "loss": 7.667045593261719,
+ "step": 37
+ },
+ {
+ "epoch": 0.5310043668122271,
+ "grad_norm": 0.036655545234680176,
+ "learning_rate": 0.00111,
+ "loss": 7.623445987701416,
+ "step": 38
+ },
+ {
+ "epoch": 0.5449781659388646,
+ "grad_norm": 0.021042725071310997,
+ "learning_rate": 0.00114,
+ "loss": 7.582821846008301,
+ "step": 39
+ },
+ {
+ "epoch": 0.5589519650655022,
+ "grad_norm": 0.022982975468039513,
+ "learning_rate": 0.00117,
+ "loss": 7.58036994934082,
+ "step": 40
+ },
+ {
+ "epoch": 0.5729257641921397,
+ "grad_norm": 0.024781864136457443,
+ "learning_rate": 0.0012000000000000001,
+ "loss": 7.5903000831604,
+ "step": 41
+ },
+ {
+ "epoch": 0.5868995633187772,
+ "grad_norm": 0.03667089343070984,
+ "learning_rate": 0.00123,
+ "loss": 7.588233947753906,
+ "step": 42
+ },
+ {
+ "epoch": 0.6008733624454149,
+ "grad_norm": 0.04083017632365227,
+ "learning_rate": 0.00126,
+ "loss": 7.548343658447266,
+ "step": 43
+ },
+ {
+ "epoch": 0.6148471615720524,
+ "grad_norm": 0.02585385926067829,
+ "learning_rate": 0.00129,
+ "loss": 7.604833602905273,
+ "step": 44
+ },
+ {
+ "epoch": 0.62882096069869,
+ "grad_norm": 0.018793931230902672,
+ "learning_rate": 0.00132,
+ "loss": 7.56705379486084,
+ "step": 45
+ },
+ {
+ "epoch": 0.6427947598253275,
+ "grad_norm": 0.025538936257362366,
+ "learning_rate": 0.00135,
+ "loss": 7.562107086181641,
+ "step": 46
+ },
+ {
+ "epoch": 0.6567685589519651,
+ "grad_norm": 0.02819318138062954,
+ "learning_rate": 0.0013800000000000002,
+ "loss": 7.517213821411133,
+ "step": 47
+ },
+ {
+ "epoch": 0.6707423580786026,
+ "grad_norm": 0.016239255666732788,
+ "learning_rate": 0.00141,
+ "loss": 7.4780755043029785,
+ "step": 48
+ },
+ {
+ "epoch": 0.6847161572052402,
+ "grad_norm": 0.024470701813697815,
+ "learning_rate": 0.0014399999999999999,
+ "loss": 7.505887985229492,
+ "step": 49
+ },
+ {
+ "epoch": 0.6986899563318777,
+ "grad_norm": 0.015499847941100597,
+ "learning_rate": 0.00147,
+ "loss": 7.484469890594482,
+ "step": 50
+ },
+ {
+ "epoch": 0.7126637554585152,
+ "grad_norm": 0.017785200849175453,
+ "learning_rate": 0.0015,
+ "loss": 7.481822967529297,
+ "step": 51
+ },
+ {
+ "epoch": 0.7266375545851529,
+ "grad_norm": 0.015952805057168007,
+ "learning_rate": 0.0015300000000000001,
+ "loss": 7.484804153442383,
+ "step": 52
+ },
+ {
+ "epoch": 0.7406113537117904,
+ "grad_norm": 0.015771260485053062,
+ "learning_rate": 0.0015600000000000002,
+ "loss": 7.4495439529418945,
+ "step": 53
+ },
+ {
+ "epoch": 0.754585152838428,
+ "grad_norm": 0.024370530620217323,
+ "learning_rate": 0.00159,
+ "loss": 7.471208572387695,
+ "step": 54
+ },
+ {
+ "epoch": 0.7685589519650655,
+ "grad_norm": 0.018173737451434135,
+ "learning_rate": 0.0016200000000000001,
+ "loss": 7.438447952270508,
+ "step": 55
+ },
+ {
+ "epoch": 0.7825327510917031,
+ "grad_norm": 0.020564094185829163,
+ "learning_rate": 0.0016500000000000002,
+ "loss": 7.437922954559326,
+ "step": 56
+ },
+ {
+ "epoch": 0.7965065502183406,
+ "grad_norm": 0.016154944896697998,
+ "learning_rate": 0.0016800000000000003,
+ "loss": 7.429378032684326,
+ "step": 57
+ },
+ {
+ "epoch": 0.8104803493449781,
+ "grad_norm": 0.016817007213830948,
+ "learning_rate": 0.00171,
+ "loss": 7.406428337097168,
+ "step": 58
+ },
+ {
+ "epoch": 0.8244541484716157,
+ "grad_norm": 0.03460947796702385,
+ "learning_rate": 0.00174,
+ "loss": 7.393331527709961,
+ "step": 59
+ },
+ {
+ "epoch": 0.8384279475982532,
+ "grad_norm": 0.15410488843917847,
+ "learning_rate": 0.0017699999999999999,
+ "loss": 7.645856857299805,
+ "step": 60
+ },
+ {
+ "epoch": 0.8524017467248908,
+ "grad_norm": 0.07443148642778397,
+ "learning_rate": 0.0018,
+ "loss": 7.549508094787598,
+ "step": 61
+ },
+ {
+ "epoch": 0.8663755458515284,
+ "grad_norm": 0.0507572740316391,
+ "learning_rate": 0.00183,
+ "loss": 7.517921447753906,
+ "step": 62
+ },
+ {
+ "epoch": 0.880349344978166,
+ "grad_norm": 0.09682746231555939,
+ "learning_rate": 0.00186,
+ "loss": 7.543978691101074,
+ "step": 63
+ },
+ {
+ "epoch": 0.8943231441048035,
+ "grad_norm": 0.2571517527103424,
+ "learning_rate": 0.00189,
+ "loss": 7.768949508666992,
+ "step": 64
+ },
+ {
+ "epoch": 0.9082969432314411,
+ "grad_norm": 0.0317784883081913,
+ "learning_rate": 0.00192,
+ "loss": 7.509063720703125,
+ "step": 65
+ },
+ {
+ "epoch": 0.9222707423580786,
+ "grad_norm": 0.0353970006108284,
+ "learning_rate": 0.0019500000000000001,
+ "loss": 7.522322654724121,
+ "step": 66
+ },
+ {
+ "epoch": 0.9362445414847161,
+ "grad_norm": 0.03663994371891022,
+ "learning_rate": 0.00198,
+ "loss": 7.504723072052002,
+ "step": 67
+ },
+ {
+ "epoch": 0.9502183406113537,
+ "grad_norm": 0.025481393560767174,
+ "learning_rate": 0.00201,
+ "loss": 7.452597141265869,
+ "step": 68
+ },
+ {
+ "epoch": 0.9641921397379912,
+ "grad_norm": 0.03099915198981762,
+ "learning_rate": 0.00204,
+ "loss": 7.455441474914551,
+ "step": 69
+ },
+ {
+ "epoch": 0.9781659388646288,
+ "grad_norm": 0.022921698167920113,
+ "learning_rate": 0.00207,
+ "loss": 7.4454827308654785,
+ "step": 70
+ },
+ {
+ "epoch": 0.9921397379912664,
+ "grad_norm": 0.032388731837272644,
+ "learning_rate": 0.0021,
+ "loss": 7.467103958129883,
+ "step": 71
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.022428609430789948,
+ "learning_rate": 0.00213,
+ "loss": 7.467540264129639,
+ "step": 72
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 7.469719409942627,
+ "eval_runtime": 50.9714,
+ "eval_samples_per_second": 47.909,
+ "eval_steps_per_second": 1.511,
+ "step": 72
+ },
+ {
+ "epoch": 1.0139737991266375,
+ "grad_norm": 0.017480911687016487,
+ "learning_rate": 0.00216,
+ "loss": 7.455810546875,
+ "step": 73
+ },
+ {
+ "epoch": 1.027947598253275,
+ "grad_norm": 0.016675131395459175,
+ "learning_rate": 0.00219,
+ "loss": 7.474848747253418,
+ "step": 74
+ },
+ {
+ "epoch": 1.0419213973799126,
+ "grad_norm": 0.017450228333473206,
+ "learning_rate": 0.00222,
+ "loss": 7.442611217498779,
+ "step": 75
+ },
+ {
+ "epoch": 1.0558951965065502,
+ "grad_norm": 0.016823675483465195,
+ "learning_rate": 0.0022500000000000003,
+ "loss": 7.454628944396973,
+ "step": 76
+ },
+ {
+ "epoch": 1.0698689956331877,
+ "grad_norm": 0.016717543825507164,
+ "learning_rate": 0.00228,
+ "loss": 7.425317764282227,
+ "step": 77
+ },
+ {
+ "epoch": 1.0838427947598253,
+ "grad_norm": 0.01644429750740528,
+ "learning_rate": 0.00231,
+ "loss": 7.4102864265441895,
+ "step": 78
+ },
+ {
+ "epoch": 1.0978165938864628,
+ "grad_norm": 0.013104400597512722,
+ "learning_rate": 0.00234,
+ "loss": 7.4559173583984375,
+ "step": 79
+ },
+ {
+ "epoch": 1.1117903930131003,
+ "grad_norm": 0.01175684668123722,
+ "learning_rate": 0.00237,
+ "loss": 7.438898086547852,
+ "step": 80
+ },
+ {
+ "epoch": 1.125764192139738,
+ "grad_norm": 0.013736684806644917,
+ "learning_rate": 0.0024000000000000002,
+ "loss": 7.449005603790283,
+ "step": 81
+ },
+ {
+ "epoch": 1.1397379912663754,
+ "grad_norm": 0.01176756713539362,
+ "learning_rate": 0.0024300000000000003,
+ "loss": 7.428421974182129,
+ "step": 82
+ },
+ {
+ "epoch": 1.1537117903930132,
+ "grad_norm": 0.011860918253660202,
+ "learning_rate": 0.00246,
+ "loss": 7.411491394042969,
+ "step": 83
+ },
+ {
+ "epoch": 1.1676855895196507,
+ "grad_norm": 0.010093477554619312,
+ "learning_rate": 0.00249,
+ "loss": 7.39311408996582,
+ "step": 84
+ },
+ {
+ "epoch": 1.1816593886462883,
+ "grad_norm": 0.009453288279473782,
+ "learning_rate": 0.00252,
+ "loss": 7.387656211853027,
+ "step": 85
+ },
+ {
+ "epoch": 1.1956331877729258,
+ "grad_norm": 0.010311692021787167,
+ "learning_rate": 0.00255,
+ "loss": 7.401076316833496,
+ "step": 86
+ },
+ {
+ "epoch": 1.2096069868995634,
+ "grad_norm": 0.010053195990622044,
+ "learning_rate": 0.00258,
+ "loss": 7.39655876159668,
+ "step": 87
+ },
+ {
+ "epoch": 1.223580786026201,
+ "grad_norm": 0.009758037514984608,
+ "learning_rate": 0.00261,
+ "loss": 7.423122882843018,
+ "step": 88
+ },
+ {
+ "epoch": 1.2375545851528384,
+ "grad_norm": 0.008394946344196796,
+ "learning_rate": 0.00264,
+ "loss": 7.383601188659668,
+ "step": 89
+ },
+ {
+ "epoch": 1.251528384279476,
+ "grad_norm": 0.008457187563180923,
+ "learning_rate": 0.00267,
+ "loss": 7.400934219360352,
+ "step": 90
+ },
+ {
+ "epoch": 1.2655021834061135,
+ "grad_norm": 0.010601334273815155,
+ "learning_rate": 0.0027,
+ "loss": 7.400241851806641,
+ "step": 91
+ },
+ {
+ "epoch": 1.279475982532751,
+ "grad_norm": 0.009009373374283314,
+ "learning_rate": 0.0027300000000000002,
+ "loss": 7.377348899841309,
+ "step": 92
+ },
+ {
+ "epoch": 1.2934497816593886,
+ "grad_norm": 0.00737508200109005,
+ "learning_rate": 0.0027600000000000003,
+ "loss": 7.385881423950195,
+ "step": 93
+ },
+ {
+ "epoch": 1.3074235807860262,
+ "grad_norm": 0.0075711836107075214,
+ "learning_rate": 0.0027900000000000004,
+ "loss": 7.348570823669434,
+ "step": 94
+ },
+ {
+ "epoch": 1.3213973799126637,
+ "grad_norm": 0.00846901722252369,
+ "learning_rate": 0.00282,
+ "loss": 7.364287853240967,
+ "step": 95
+ },
+ {
+ "epoch": 1.3353711790393012,
+ "grad_norm": 0.007160635199397802,
+ "learning_rate": 0.00285,
+ "loss": 7.3881635665893555,
+ "step": 96
+ },
+ {
+ "epoch": 1.3493449781659388,
+ "grad_norm": 0.00682067358866334,
+ "learning_rate": 0.0028799999999999997,
+ "loss": 7.348663330078125,
+ "step": 97
+ },
+ {
+ "epoch": 1.3633187772925766,
+ "grad_norm": 0.009108101949095726,
+ "learning_rate": 0.00291,
+ "loss": 7.356302261352539,
+ "step": 98
+ },
+ {
+ "epoch": 1.3772925764192139,
+ "grad_norm": 0.015828927978873253,
+ "learning_rate": 0.00294,
+ "loss": 7.347576141357422,
+ "step": 99
+ },
+ {
+ "epoch": 1.3912663755458516,
+ "grad_norm": 0.009683096781373024,
+ "learning_rate": 0.00297,
+ "loss": 7.323967933654785,
+ "step": 100
+ },
+ {
+ "epoch": 1.405240174672489,
+ "grad_norm": 0.010833270847797394,
+ "learning_rate": 0.003,
+ "loss": 7.3139495849609375,
+ "step": 101
+ },
+ {
+ "epoch": 1.4192139737991267,
+ "grad_norm": 0.014905400574207306,
+ "learning_rate": 0.002999999401781161,
+ "loss": 7.302030086517334,
+ "step": 102
+ },
+ {
+ "epoch": 1.4331877729257643,
+ "grad_norm": 0.023674853146076202,
+ "learning_rate": 0.002999997607125127,
+ "loss": 7.311811923980713,
+ "step": 103
+ },
+ {
+ "epoch": 1.4471615720524018,
+ "grad_norm": 0.050460800528526306,
+ "learning_rate": 0.002999994616033342,
+ "loss": 7.41972541809082,
+ "step": 104
+ },
+ {
+ "epoch": 1.4611353711790394,
+ "grad_norm": 0.02193174883723259,
+ "learning_rate": 0.002999990428508218,
+ "loss": 7.294468879699707,
+ "step": 105
+ },
+ {
+ "epoch": 1.475109170305677,
+ "grad_norm": 0.036261629313230515,
+ "learning_rate": 0.0029999850445531283,
+ "loss": 7.295568466186523,
+ "step": 106
+ },
+ {
+ "epoch": 1.4890829694323144,
+ "grad_norm": 0.02734929881989956,
+ "learning_rate": 0.002999978464172409,
+ "loss": 7.309239864349365,
+ "step": 107
+ },
+ {
+ "epoch": 1.503056768558952,
+ "grad_norm": 0.017884107306599617,
+ "learning_rate": 0.002999970687371363,
+ "loss": 7.2427592277526855,
+ "step": 108
+ },
+ {
+ "epoch": 1.5170305676855895,
+ "grad_norm": 0.01700308918952942,
+ "learning_rate": 0.0029999617141562573,
+ "loss": 7.241730213165283,
+ "step": 109
+ },
+ {
+ "epoch": 1.531004366812227,
+ "grad_norm": 0.0209935512393713,
+ "learning_rate": 0.0029999515445343194,
+ "loss": 7.262085437774658,
+ "step": 110
+ },
+ {
+ "epoch": 1.5449781659388646,
+ "grad_norm": 0.023207105696201324,
+ "learning_rate": 0.0029999401785137437,
+ "loss": 7.25450325012207,
+ "step": 111
+ },
+ {
+ "epoch": 1.5589519650655022,
+ "grad_norm": 0.01907353848218918,
+ "learning_rate": 0.002999927616103687,
+ "loss": 7.23149299621582,
+ "step": 112
+ },
+ {
+ "epoch": 1.5729257641921397,
+ "grad_norm": 0.016149666160345078,
+ "learning_rate": 0.0029999138573142715,
+ "loss": 7.210232734680176,
+ "step": 113
+ },
+ {
+ "epoch": 1.5868995633187772,
+ "grad_norm": 0.017513928934931755,
+ "learning_rate": 0.002999898902156582,
+ "loss": 7.218451976776123,
+ "step": 114
+ },
+ {
+ "epoch": 1.600873362445415,
+ "grad_norm": 0.015002547763288021,
+ "learning_rate": 0.002999882750642667,
+ "loss": 7.169316291809082,
+ "step": 115
+ },
+ {
+ "epoch": 1.6148471615720523,
+ "grad_norm": 0.014233081601560116,
+ "learning_rate": 0.0029998654027855406,
+ "loss": 7.194825172424316,
+ "step": 116
+ },
+ {
+ "epoch": 1.62882096069869,
+ "grad_norm": 0.01283260714262724,
+ "learning_rate": 0.0029998468585991783,
+ "loss": 7.15669059753418,
+ "step": 117
+ },
+ {
+ "epoch": 1.6427947598253274,
+ "grad_norm": 0.014416799880564213,
+ "learning_rate": 0.0029998271180985223,
+ "loss": 7.143369674682617,
+ "step": 118
+ },
+ {
+ "epoch": 1.6567685589519652,
+ "grad_norm": 0.014337058179080486,
+ "learning_rate": 0.0029998061812994757,
+ "loss": 7.127734184265137,
+ "step": 119
+ },
+ {
+ "epoch": 1.6707423580786025,
+ "grad_norm": 0.016406210139393806,
+ "learning_rate": 0.002999784048218908,
+ "loss": 7.086504936218262,
+ "step": 120
+ },
+ {
+ "epoch": 1.6847161572052403,
+ "grad_norm": 0.01680421642959118,
+ "learning_rate": 0.002999760718874652,
+ "loss": 7.089420318603516,
+ "step": 121
+ },
+ {
+ "epoch": 1.6986899563318776,
+ "grad_norm": 0.010665548034012318,
+ "learning_rate": 0.0029997361932855016,
+ "loss": 7.102463722229004,
+ "step": 122
+ },
+ {
+ "epoch": 1.7126637554585153,
+ "grad_norm": 0.013479137793183327,
+ "learning_rate": 0.002999710471471218,
+ "loss": 7.079098701477051,
+ "step": 123
+ },
+ {
+ "epoch": 1.726637554585153,
+ "grad_norm": 0.012408486567437649,
+ "learning_rate": 0.0029996835534525246,
+ "loss": 7.0790114402771,
+ "step": 124
+ },
+ {
+ "epoch": 1.7406113537117904,
+ "grad_norm": 0.012035196647047997,
+ "learning_rate": 0.0029996554392511094,
+ "loss": 7.058756351470947,
+ "step": 125
+ },
+ {
+ "epoch": 1.754585152838428,
+ "grad_norm": 0.00899985246360302,
+ "learning_rate": 0.0029996261288896225,
+ "loss": 7.0739240646362305,
+ "step": 126
+ },
+ {
+ "epoch": 1.7685589519650655,
+ "grad_norm": 0.01072906143963337,
+ "learning_rate": 0.002999595622391679,
+ "loss": 7.085537910461426,
+ "step": 127
+ },
+ {
+ "epoch": 1.782532751091703,
+ "grad_norm": 0.01047342736274004,
+ "learning_rate": 0.0029995639197818577,
+ "loss": 7.060049057006836,
+ "step": 128
+ },
+ {
+ "epoch": 1.7965065502183406,
+ "grad_norm": 0.02342618815600872,
+ "learning_rate": 0.0029995310210857006,
+ "loss": 7.039482116699219,
+ "step": 129
+ },
+ {
+ "epoch": 1.8104803493449781,
+ "grad_norm": 0.011256859637796879,
+ "learning_rate": 0.0029994969263297134,
+ "loss": 7.040412902832031,
+ "step": 130
+ },
+ {
+ "epoch": 1.8244541484716157,
+ "grad_norm": 0.008407019078731537,
+ "learning_rate": 0.0029994616355413666,
+ "loss": 7.0364274978637695,
+ "step": 131
+ },
+ {
+ "epoch": 1.8384279475982532,
+ "grad_norm": 0.011006148532032967,
+ "learning_rate": 0.0029994251487490916,
+ "loss": 6.991468906402588,
+ "step": 132
+ },
+ {
+ "epoch": 1.8524017467248908,
+ "grad_norm": 0.009103255346417427,
+ "learning_rate": 0.0029993874659822875,
+ "loss": 7.001059532165527,
+ "step": 133
+ },
+ {
+ "epoch": 1.8663755458515285,
+ "grad_norm": 0.007875788025557995,
+ "learning_rate": 0.0029993485872713125,
+ "loss": 6.9974260330200195,
+ "step": 134
+ },
+ {
+ "epoch": 1.8803493449781659,
+ "grad_norm": 0.007872611284255981,
+ "learning_rate": 0.0029993085126474913,
+ "loss": 6.980216979980469,
+ "step": 135
+ },
+ {
+ "epoch": 1.8943231441048036,
+ "grad_norm": 0.010071957483887672,
+ "learning_rate": 0.0029992672421431115,
+ "loss": 6.97799015045166,
+ "step": 136
+ },
+ {
+ "epoch": 1.908296943231441,
+ "grad_norm": 0.01209618803113699,
+ "learning_rate": 0.0029992247757914237,
+ "loss": 6.97202205657959,
+ "step": 137
+ },
+ {
+ "epoch": 1.9222707423580787,
+ "grad_norm": 0.014884858392179012,
+ "learning_rate": 0.0029991811136266425,
+ "loss": 6.976570129394531,
+ "step": 138
+ },
+ {
+ "epoch": 1.936244541484716,
+ "grad_norm": 0.014953463338315487,
+ "learning_rate": 0.0029991362556839453,
+ "loss": 6.971050262451172,
+ "step": 139
+ },
+ {
+ "epoch": 1.9502183406113538,
+ "grad_norm": 0.008371063508093357,
+ "learning_rate": 0.002999090201999474,
+ "loss": 6.955556869506836,
+ "step": 140
+ },
+ {
+ "epoch": 1.9641921397379911,
+ "grad_norm": 0.011926881037652493,
+ "learning_rate": 0.002999042952610332,
+ "loss": 6.943947792053223,
+ "step": 141
+ },
+ {
+ "epoch": 1.9781659388646289,
+ "grad_norm": 0.013422893360257149,
+ "learning_rate": 0.002998994507554589,
+ "loss": 6.910022735595703,
+ "step": 142
+ },
+ {
+ "epoch": 1.9921397379912664,
+ "grad_norm": 0.009672224521636963,
+ "learning_rate": 0.002998944866871275,
+ "loss": 6.8994855880737305,
+ "step": 143
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.012590581551194191,
+ "learning_rate": 0.0029988940306003853,
+ "loss": 6.88709020614624,
+ "step": 144
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 6.945146083831787,
+ "eval_runtime": 51.5801,
+ "eval_samples_per_second": 47.344,
+ "eval_steps_per_second": 1.493,
+ "step": 144
+ },
+ {
+ "epoch": 2.0139737991266378,
+ "grad_norm": 0.016554836183786392,
+ "learning_rate": 0.002998841998782877,
+ "loss": 6.916609764099121,
+ "step": 145
+ },
+ {
+ "epoch": 2.027947598253275,
+ "grad_norm": 0.011085660196840763,
+ "learning_rate": 0.0029987887714606723,
+ "loss": 6.884910583496094,
+ "step": 146
+ },
+ {
+ "epoch": 2.041921397379913,
+ "grad_norm": 0.008596168830990791,
+ "learning_rate": 0.0029987343486766547,
+ "loss": 6.900066375732422,
+ "step": 147
+ },
+ {
+ "epoch": 2.05589519650655,
+ "grad_norm": 0.014487302862107754,
+ "learning_rate": 0.0029986787304746715,
+ "loss": 6.8693342208862305,
+ "step": 148
+ },
+ {
+ "epoch": 2.069868995633188,
+ "grad_norm": 0.015636997297406197,
+ "learning_rate": 0.002998621916899534,
+ "loss": 6.903670787811279,
+ "step": 149
+ },
+ {
+ "epoch": 2.0838427947598253,
+ "grad_norm": 0.013275462202727795,
+ "learning_rate": 0.0029985639079970153,
+ "loss": 6.889361381530762,
+ "step": 150
+ },
+ {
+ "epoch": 2.097816593886463,
+ "grad_norm": 0.009044744074344635,
+ "learning_rate": 0.002998504703813852,
+ "loss": 6.8549981117248535,
+ "step": 151
+ },
+ {
+ "epoch": 2.1117903930131003,
+ "grad_norm": 0.011327696964144707,
+ "learning_rate": 0.002998444304397745,
+ "loss": 6.856151103973389,
+ "step": 152
+ },
+ {
+ "epoch": 2.125764192139738,
+ "grad_norm": 0.01310331467539072,
+ "learning_rate": 0.0029983827097973554,
+ "loss": 6.834944725036621,
+ "step": 153
+ },
+ {
+ "epoch": 2.1397379912663754,
+ "grad_norm": 0.010003753937780857,
+ "learning_rate": 0.00299831992006231,
+ "loss": 6.825613975524902,
+ "step": 154
+ },
+ {
+ "epoch": 2.153711790393013,
+ "grad_norm": 0.009928115643560886,
+ "learning_rate": 0.0029982559352431967,
+ "loss": 6.828935146331787,
+ "step": 155
+ },
+ {
+ "epoch": 2.1676855895196505,
+ "grad_norm": 0.012302150949835777,
+ "learning_rate": 0.0029981907553915674,
+ "loss": 6.821207046508789,
+ "step": 156
+ },
+ {
+ "epoch": 2.1816593886462883,
+ "grad_norm": 0.020555168390274048,
+ "learning_rate": 0.002998124380559936,
+ "loss": 6.8184895515441895,
+ "step": 157
+ },
+ {
+ "epoch": 2.1956331877729256,
+ "grad_norm": 0.011417638510465622,
+ "learning_rate": 0.0029980568108017797,
+ "loss": 6.788754463195801,
+ "step": 158
+ },
+ {
+ "epoch": 2.2096069868995634,
+ "grad_norm": 0.010772978886961937,
+ "learning_rate": 0.002997988046171538,
+ "loss": 6.7861433029174805,
+ "step": 159
+ },
+ {
+ "epoch": 2.2235807860262007,
+ "grad_norm": 0.012753859162330627,
+ "learning_rate": 0.002997918086724614,
+ "loss": 6.788776397705078,
+ "step": 160
+ },
+ {
+ "epoch": 2.2375545851528384,
+ "grad_norm": 0.011327114887535572,
+ "learning_rate": 0.0029978469325173717,
+ "loss": 6.779160022735596,
+ "step": 161
+ },
+ {
+ "epoch": 2.251528384279476,
+ "grad_norm": 0.01235753484070301,
+ "learning_rate": 0.002997774583607139,
+ "loss": 6.799554824829102,
+ "step": 162
+ },
+ {
+ "epoch": 2.2655021834061135,
+ "grad_norm": 0.01417156495153904,
+ "learning_rate": 0.0029977010400522074,
+ "loss": 6.77957820892334,
+ "step": 163
+ },
+ {
+ "epoch": 2.279475982532751,
+ "grad_norm": 0.021321553736925125,
+ "learning_rate": 0.002997626301911828,
+ "loss": 6.785632133483887,
+ "step": 164
+ },
+ {
+ "epoch": 2.2934497816593886,
+ "grad_norm": 0.03614472970366478,
+ "learning_rate": 0.002997550369246217,
+ "loss": 6.836414337158203,
+ "step": 165
+ },
+ {
+ "epoch": 2.3074235807860264,
+ "grad_norm": 0.03934052586555481,
+ "learning_rate": 0.002997473242116551,
+ "loss": 6.854659557342529,
+ "step": 166
+ },
+ {
+ "epoch": 2.3213973799126637,
+ "grad_norm": 0.016240060329437256,
+ "learning_rate": 0.0029973949205849707,
+ "loss": 6.789690971374512,
+ "step": 167
+ },
+ {
+ "epoch": 2.3353711790393015,
+ "grad_norm": 0.031164905056357384,
+ "learning_rate": 0.0029973154047145782,
+ "loss": 6.8363776206970215,
+ "step": 168
+ },
+ {
+ "epoch": 2.349344978165939,
+ "grad_norm": 0.013051844201982021,
+ "learning_rate": 0.002997234694569438,
+ "loss": 6.803576469421387,
+ "step": 169
+ },
+ {
+ "epoch": 2.3633187772925766,
+ "grad_norm": 0.022611448541283607,
+ "learning_rate": 0.002997152790214577,
+ "loss": 6.762107849121094,
+ "step": 170
+ },
+ {
+ "epoch": 2.377292576419214,
+ "grad_norm": 0.017812907695770264,
+ "learning_rate": 0.0029970696917159833,
+ "loss": 6.776218414306641,
+ "step": 171
+ },
+ {
+ "epoch": 2.3912663755458516,
+ "grad_norm": 0.015453332103788853,
+ "learning_rate": 0.0029969853991406088,
+ "loss": 6.742618083953857,
+ "step": 172
+ },
+ {
+ "epoch": 2.405240174672489,
+ "grad_norm": 0.020630091428756714,
+ "learning_rate": 0.0029968999125563655,
+ "loss": 6.765475749969482,
+ "step": 173
+ },
+ {
+ "epoch": 2.4192139737991267,
+ "grad_norm": 0.014672341756522655,
+ "learning_rate": 0.0029968132320321297,
+ "loss": 6.754009246826172,
+ "step": 174
+ },
+ {
+ "epoch": 2.433187772925764,
+ "grad_norm": 0.016540220007300377,
+ "learning_rate": 0.0029967253576377376,
+ "loss": 6.734752655029297,
+ "step": 175
+ },
+ {
+ "epoch": 2.447161572052402,
+ "grad_norm": 0.013495649211108685,
+ "learning_rate": 0.0029966362894439874,
+ "loss": 6.752676010131836,
+ "step": 176
+ },
+ {
+ "epoch": 2.461135371179039,
+ "grad_norm": 0.015936989337205887,
+ "learning_rate": 0.0029965460275226405,
+ "loss": 6.725708961486816,
+ "step": 177
+ },
+ {
+ "epoch": 2.475109170305677,
+ "grad_norm": 0.010431243106722832,
+ "learning_rate": 0.0029964545719464194,
+ "loss": 6.698376655578613,
+ "step": 178
+ },
+ {
+ "epoch": 2.489082969432314,
+ "grad_norm": 0.01669321395456791,
+ "learning_rate": 0.0029963619227890083,
+ "loss": 6.711770534515381,
+ "step": 179
+ },
+ {
+ "epoch": 2.503056768558952,
+ "grad_norm": 0.013308997265994549,
+ "learning_rate": 0.0029962680801250514,
+ "loss": 6.705286979675293,
+ "step": 180
+ },
+ {
+ "epoch": 2.5170305676855893,
+ "grad_norm": 0.011648476123809814,
+ "learning_rate": 0.002996173044030159,
+ "loss": 6.690052509307861,
+ "step": 181
+ },
+ {
+ "epoch": 2.531004366812227,
+ "grad_norm": 0.03942161798477173,
+ "learning_rate": 0.0029960768145808965,
+ "loss": 6.711885452270508,
+ "step": 182
+ },
+ {
+ "epoch": 2.544978165938865,
+ "grad_norm": 0.01571274921298027,
+ "learning_rate": 0.002995979391854797,
+ "loss": 6.715847969055176,
+ "step": 183
+ },
+ {
+ "epoch": 2.558951965065502,
+ "grad_norm": 0.00829714722931385,
+ "learning_rate": 0.0029958807759303515,
+ "loss": 6.685308933258057,
+ "step": 184
+ },
+ {
+ "epoch": 2.5729257641921395,
+ "grad_norm": 0.01374727487564087,
+ "learning_rate": 0.0029957809668870114,
+ "loss": 6.695744514465332,
+ "step": 185
+ },
+ {
+ "epoch": 2.5868995633187772,
+ "grad_norm": 0.010857059620320797,
+ "learning_rate": 0.0029956799648051934,
+ "loss": 6.658294677734375,
+ "step": 186
+ },
+ {
+ "epoch": 2.600873362445415,
+ "grad_norm": 0.008897246792912483,
+ "learning_rate": 0.002995577769766272,
+ "loss": 6.642439842224121,
+ "step": 187
+ },
+ {
+ "epoch": 2.6148471615720523,
+ "grad_norm": 0.015880772843956947,
+ "learning_rate": 0.002995474381852584,
+ "loss": 6.674907684326172,
+ "step": 188
+ },
+ {
+ "epoch": 2.62882096069869,
+ "grad_norm": 0.021623898297548294,
+ "learning_rate": 0.0029953698011474265,
+ "loss": 6.66236686706543,
+ "step": 189
+ },
+ {
+ "epoch": 2.6427947598253274,
+ "grad_norm": 0.03084627166390419,
+ "learning_rate": 0.0029952640277350596,
+ "loss": 6.666715621948242,
+ "step": 190
+ },
+ {
+ "epoch": 2.656768558951965,
+ "grad_norm": 0.040561504662036896,
+ "learning_rate": 0.002995157061700702,
+ "loss": 6.649062156677246,
+ "step": 191
+ },
+ {
+ "epoch": 2.6707423580786025,
+ "grad_norm": 0.031126515939831734,
+ "learning_rate": 0.0029950489031305345,
+ "loss": 6.640542030334473,
+ "step": 192
+ },
+ {
+ "epoch": 2.6847161572052403,
+ "grad_norm": 0.00872360821813345,
+ "learning_rate": 0.0029949395521116987,
+ "loss": 6.635221481323242,
+ "step": 193
+ },
+ {
+ "epoch": 2.6986899563318776,
+ "grad_norm": 0.023432321846485138,
+ "learning_rate": 0.0029948290087322966,
+ "loss": 6.642196178436279,
+ "step": 194
+ },
+ {
+ "epoch": 2.7126637554585153,
+ "grad_norm": 0.021096013486385345,
+ "learning_rate": 0.0029947172730813913,
+ "loss": 6.643655300140381,
+ "step": 195
+ },
+ {
+ "epoch": 2.726637554585153,
+ "grad_norm": 0.014494858682155609,
+ "learning_rate": 0.002994604345249006,
+ "loss": 6.612720966339111,
+ "step": 196
+ },
+ {
+ "epoch": 2.7406113537117904,
+ "grad_norm": 0.05225599184632301,
+ "learning_rate": 0.0029944902253261248,
+ "loss": 6.60659646987915,
+ "step": 197
+ },
+ {
+ "epoch": 2.7545851528384278,
+ "grad_norm": 0.014311269856989384,
+ "learning_rate": 0.0029943749134046915,
+ "loss": 6.54373025894165,
+ "step": 198
+ },
+ {
+ "epoch": 2.7685589519650655,
+ "grad_norm": 0.010777472518384457,
+ "learning_rate": 0.0029942584095776115,
+ "loss": 6.599837303161621,
+ "step": 199
+ },
+ {
+ "epoch": 2.7825327510917033,
+ "grad_norm": 0.012498079799115658,
+ "learning_rate": 0.002994140713938749,
+ "loss": 6.589657783508301,
+ "step": 200
+ },
+ {
+ "epoch": 2.7965065502183406,
+ "grad_norm": 0.015775132924318314,
+ "learning_rate": 0.00299402182658293,
+ "loss": 6.577511787414551,
+ "step": 201
+ },
+ {
+ "epoch": 2.810480349344978,
+ "grad_norm": 0.017978230491280556,
+ "learning_rate": 0.0029939017476059396,
+ "loss": 6.58717155456543,
+ "step": 202
+ },
+ {
+ "epoch": 2.8244541484716157,
+ "grad_norm": 0.026807745918631554,
+ "learning_rate": 0.0029937804771045236,
+ "loss": 6.578145503997803,
+ "step": 203
+ },
+ {
+ "epoch": 2.8384279475982535,
+ "grad_norm": 0.04853282868862152,
+ "learning_rate": 0.0029936580151763872,
+ "loss": 6.609755516052246,
+ "step": 204
+ },
+ {
+ "epoch": 2.8524017467248908,
+ "grad_norm": 0.1057916209101677,
+ "learning_rate": 0.002993534361920195,
+ "loss": 6.870367050170898,
+ "step": 205
+ },
+ {
+ "epoch": 2.8663755458515285,
+ "grad_norm": 0.04925911873579025,
+ "learning_rate": 0.002993409517435574,
+ "loss": 6.871556282043457,
+ "step": 206
+ },
+ {
+ "epoch": 2.880349344978166,
+ "grad_norm": 0.05288556218147278,
+ "learning_rate": 0.0029932834818231077,
+ "loss": 6.808315753936768,
+ "step": 207
+ },
+ {
+ "epoch": 2.8943231441048036,
+ "grad_norm": 0.03831710293889046,
+ "learning_rate": 0.0029931562551843405,
+ "loss": 6.827617645263672,
+ "step": 208
+ },
+ {
+ "epoch": 2.908296943231441,
+ "grad_norm": 0.03665117174386978,
+ "learning_rate": 0.002993027837621778,
+ "loss": 6.767284870147705,
+ "step": 209
+ },
+ {
+ "epoch": 2.9222707423580787,
+ "grad_norm": 0.026216736063361168,
+ "learning_rate": 0.0029928982292388826,
+ "loss": 6.733659744262695,
+ "step": 210
+ },
+ {
+ "epoch": 2.936244541484716,
+ "grad_norm": 0.024796854704618454,
+ "learning_rate": 0.002992767430140078,
+ "loss": 6.759577751159668,
+ "step": 211
+ },
+ {
+ "epoch": 2.950218340611354,
+ "grad_norm": 0.02410992607474327,
+ "learning_rate": 0.002992635440430747,
+ "loss": 6.701825141906738,
+ "step": 212
+ },
+ {
+ "epoch": 2.964192139737991,
+ "grad_norm": 0.019924134016036987,
+ "learning_rate": 0.0029925022602172308,
+ "loss": 6.677121162414551,
+ "step": 213
+ },
+ {
+ "epoch": 2.978165938864629,
+ "grad_norm": 0.022516867145895958,
+ "learning_rate": 0.00299236788960683,
+ "loss": 6.668622970581055,
+ "step": 214
+ },
+ {
+ "epoch": 2.992139737991266,
+ "grad_norm": 0.025068556889891624,
+ "learning_rate": 0.002992232328707806,
+ "loss": 6.67718505859375,
+ "step": 215
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.02574208751320839,
+ "learning_rate": 0.002992095577629376,
+ "loss": 6.679594039916992,
+ "step": 216
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 6.657555103302002,
+ "eval_runtime": 51.0199,
+ "eval_samples_per_second": 47.864,
+ "eval_steps_per_second": 1.509,
+ "step": 216
+ },
+ {
+ "epoch": 3.0139737991266378,
+ "grad_norm": 0.017298126593232155,
+ "learning_rate": 0.0029919576364817184,
+ "loss": 6.620774269104004,
+ "step": 217
+ },
+ {
+ "epoch": 3.027947598253275,
+ "grad_norm": 0.018840325996279716,
+ "learning_rate": 0.002991818505375971,
+ "loss": 6.613316535949707,
+ "step": 218
+ },
+ {
+ "epoch": 3.041921397379913,
+ "grad_norm": 0.020813113078475,
+ "learning_rate": 0.002991678184424228,
+ "loss": 6.621555328369141,
+ "step": 219
+ },
+ {
+ "epoch": 3.05589519650655,
+ "grad_norm": 0.026581479236483574,
+ "learning_rate": 0.0029915366737395437,
+ "loss": 6.613471031188965,
+ "step": 220
+ },
+ {
+ "epoch": 3.069868995633188,
+ "grad_norm": 0.02084030583500862,
+ "learning_rate": 0.0029913939734359306,
+ "loss": 6.559704780578613,
+ "step": 221
+ },
+ {
+ "epoch": 3.0838427947598253,
+ "grad_norm": 0.01327502354979515,
+ "learning_rate": 0.00299125008362836,
+ "loss": 6.564910888671875,
+ "step": 222
+ },
+ {
+ "epoch": 3.097816593886463,
+ "grad_norm": 0.015091734007000923,
+ "learning_rate": 0.0029911050044327614,
+ "loss": 6.571262836456299,
+ "step": 223
+ },
+ {
+ "epoch": 3.1117903930131003,
+ "grad_norm": 0.015463982708752155,
+ "learning_rate": 0.002990958735966023,
+ "loss": 6.565750598907471,
+ "step": 224
+ },
+ {
+ "epoch": 3.125764192139738,
+ "grad_norm": 0.013636826537549496,
+ "learning_rate": 0.0029908112783459894,
+ "loss": 6.57981014251709,
+ "step": 225
+ },
+ {
+ "epoch": 3.1397379912663754,
+ "grad_norm": 0.010613273829221725,
+ "learning_rate": 0.0029906626316914655,
+ "loss": 6.538505554199219,
+ "step": 226
+ },
+ {
+ "epoch": 3.153711790393013,
+ "grad_norm": 0.01804237812757492,
+ "learning_rate": 0.0029905127961222127,
+ "loss": 6.5282511711120605,
+ "step": 227
+ },
+ {
+ "epoch": 3.1676855895196505,
+ "grad_norm": 0.015131509862840176,
+ "learning_rate": 0.0029903617717589506,
+ "loss": 6.531335830688477,
+ "step": 228
+ },
+ {
+ "epoch": 3.1816593886462883,
+ "grad_norm": 0.010518206283450127,
+ "learning_rate": 0.002990209558723358,
+ "loss": 6.521886825561523,
+ "step": 229
+ },
+ {
+ "epoch": 3.1956331877729256,
+ "grad_norm": 0.011360022239387035,
+ "learning_rate": 0.002990056157138069,
+ "loss": 6.528301239013672,
+ "step": 230
+ },
+ {
+ "epoch": 3.2096069868995634,
+ "grad_norm": 0.011556181125342846,
+ "learning_rate": 0.0029899015671266773,
+ "loss": 6.497917652130127,
+ "step": 231
+ },
+ {
+ "epoch": 3.2235807860262007,
+ "grad_norm": 0.01365791168063879,
+ "learning_rate": 0.0029897457888137325,
+ "loss": 6.52849006652832,
+ "step": 232
+ },
+ {
+ "epoch": 3.2375545851528384,
+ "grad_norm": 0.01091043371707201,
+ "learning_rate": 0.0029895888223247426,
+ "loss": 6.492197036743164,
+ "step": 233
+ },
+ {
+ "epoch": 3.251528384279476,
+ "grad_norm": 0.008484847843647003,
+ "learning_rate": 0.002989430667786173,
+ "loss": 6.497377395629883,
+ "step": 234
+ },
+ {
+ "epoch": 3.2655021834061135,
+ "grad_norm": 0.00836879201233387,
+ "learning_rate": 0.0029892713253254466,
+ "loss": 6.495189666748047,
+ "step": 235
+ },
+ {
+ "epoch": 3.279475982532751,
+ "grad_norm": 0.009781497530639172,
+ "learning_rate": 0.0029891107950709406,
+ "loss": 6.454085826873779,
+ "step": 236
+ },
+ {
+ "epoch": 3.2934497816593886,
+ "grad_norm": 0.009603648446500301,
+ "learning_rate": 0.0029889490771519934,
+ "loss": 6.499475479125977,
+ "step": 237
+ },
+ {
+ "epoch": 3.3074235807860264,
+ "grad_norm": 0.007991855964064598,
+ "learning_rate": 0.0029887861716988975,
+ "loss": 6.460785865783691,
+ "step": 238
+ },
+ {
+ "epoch": 3.3213973799126637,
+ "grad_norm": 0.008447187952697277,
+ "learning_rate": 0.0029886220788429025,
+ "loss": 6.504582405090332,
+ "step": 239
+ },
+ {
+ "epoch": 3.3353711790393015,
+ "grad_norm": 0.010945498012006283,
+ "learning_rate": 0.0029884567987162156,
+ "loss": 6.49589729309082,
+ "step": 240
+ },
+ {
+ "epoch": 3.349344978165939,
+ "grad_norm": 0.011096215806901455,
+ "learning_rate": 0.0029882903314519997,
+ "loss": 6.460657119750977,
+ "step": 241
+ },
+ {
+ "epoch": 3.3633187772925766,
+ "grad_norm": 0.013094739057123661,
+ "learning_rate": 0.0029881226771843747,
+ "loss": 6.462450981140137,
+ "step": 242
+ },
+ {
+ "epoch": 3.377292576419214,
+ "grad_norm": 0.015423037111759186,
+ "learning_rate": 0.0029879538360484165,
+ "loss": 6.45565128326416,
+ "step": 243
+ },
+ {
+ "epoch": 3.3912663755458516,
+ "grad_norm": 0.0182222630828619,
+ "learning_rate": 0.002987783808180157,
+ "loss": 6.452227592468262,
+ "step": 244
+ },
+ {
+ "epoch": 3.405240174672489,
+ "grad_norm": 0.01981227844953537,
+ "learning_rate": 0.002987612593716585,
+ "loss": 6.477701663970947,
+ "step": 245
+ },
+ {
+ "epoch": 3.4192139737991267,
+ "grad_norm": 0.026733700186014175,
+ "learning_rate": 0.002987440192795645,
+ "loss": 6.452212333679199,
+ "step": 246
+ },
+ {
+ "epoch": 3.433187772925764,
+ "grad_norm": 0.03504155948758125,
+ "learning_rate": 0.0029872666055562372,
+ "loss": 6.509665489196777,
+ "step": 247
+ },
+ {
+ "epoch": 3.447161572052402,
+ "grad_norm": 0.027970125898718834,
+ "learning_rate": 0.002987091832138217,
+ "loss": 6.487089157104492,
+ "step": 248
+ },
+ {
+ "epoch": 3.461135371179039,
+ "grad_norm": 0.013714591972529888,
+ "learning_rate": 0.0029869158726823974,
+ "loss": 6.469569206237793,
+ "step": 249
+ },
+ {
+ "epoch": 3.475109170305677,
+ "grad_norm": 0.016472695395350456,
+ "learning_rate": 0.0029867387273305453,
+ "loss": 6.461688995361328,
+ "step": 250
+ },
+ {
+ "epoch": 3.489082969432314,
+ "grad_norm": 0.018457358703017235,
+ "learning_rate": 0.0029865603962253833,
+ "loss": 6.490985870361328,
+ "step": 251
+ },
+ {
+ "epoch": 3.503056768558952,
+ "grad_norm": 0.011045100167393684,
+ "learning_rate": 0.0029863808795105896,
+ "loss": 6.403701305389404,
+ "step": 252
+ },
+ {
+ "epoch": 3.5170305676855893,
+ "grad_norm": 0.014126897789537907,
+ "learning_rate": 0.0029862001773307972,
+ "loss": 6.379804611206055,
+ "step": 253
+ },
+ {
+ "epoch": 3.531004366812227,
+ "grad_norm": 0.014015335589647293,
+ "learning_rate": 0.0029860182898315954,
+ "loss": 6.404377460479736,
+ "step": 254
+ },
+ {
+ "epoch": 3.544978165938865,
+ "grad_norm": 0.010032658465206623,
+ "learning_rate": 0.002985835217159527,
+ "loss": 6.40291690826416,
+ "step": 255
+ },
+ {
+ "epoch": 3.558951965065502,
+ "grad_norm": 0.015469838865101337,
+ "learning_rate": 0.0029856509594620912,
+ "loss": 6.4414520263671875,
+ "step": 256
+ },
+ {
+ "epoch": 3.5729257641921395,
+ "grad_norm": 0.011106910184025764,
+ "learning_rate": 0.0029854655168877397,
+ "loss": 6.396444320678711,
+ "step": 257
+ },
+ {
+ "epoch": 3.5868995633187772,
+ "grad_norm": 0.01262320950627327,
+ "learning_rate": 0.0029852788895858816,
+ "loss": 6.385222434997559,
+ "step": 258
+ },
+ {
+ "epoch": 3.600873362445415,
+ "grad_norm": 0.012783976271748543,
+ "learning_rate": 0.002985091077706878,
+ "loss": 6.3818135261535645,
+ "step": 259
+ },
+ {
+ "epoch": 3.6148471615720523,
+ "grad_norm": 0.01224067248404026,
+ "learning_rate": 0.0029849020814020468,
+ "loss": 6.37886905670166,
+ "step": 260
+ },
+ {
+ "epoch": 3.62882096069869,
+ "grad_norm": 0.010540086776018143,
+ "learning_rate": 0.0029847119008236576,
+ "loss": 6.37799072265625,
+ "step": 261
+ },
+ {
+ "epoch": 3.6427947598253274,
+ "grad_norm": 0.012479913420975208,
+ "learning_rate": 0.0029845205361249366,
+ "loss": 6.365094184875488,
+ "step": 262
+ },
+ {
+ "epoch": 3.656768558951965,
+ "grad_norm": 0.010080077685415745,
+ "learning_rate": 0.0029843279874600622,
+ "loss": 6.371852874755859,
+ "step": 263
+ },
+ {
+ "epoch": 3.6707423580786025,
+ "grad_norm": 0.013266298919916153,
+ "learning_rate": 0.002984134254984168,
+ "loss": 6.380039215087891,
+ "step": 264
+ },
+ {
+ "epoch": 3.6847161572052403,
+ "grad_norm": 0.01014771405607462,
+ "learning_rate": 0.002983939338853339,
+ "loss": 6.370337963104248,
+ "step": 265
+ },
+ {
+ "epoch": 3.6986899563318776,
+ "grad_norm": 0.008164579048752785,
+ "learning_rate": 0.0029837432392246185,
+ "loss": 6.331899642944336,
+ "step": 266
+ },
+ {
+ "epoch": 3.7126637554585153,
+ "grad_norm": 0.01020442508161068,
+ "learning_rate": 0.002983545956255998,
+ "loss": 6.35077428817749,
+ "step": 267
+ },
+ {
+ "epoch": 3.726637554585153,
+ "grad_norm": 0.009898579679429531,
+ "learning_rate": 0.002983347490106426,
+ "loss": 6.371985912322998,
+ "step": 268
+ },
+ {
+ "epoch": 3.7406113537117904,
+ "grad_norm": 0.009668919257819653,
+ "learning_rate": 0.002983147840935802,
+ "loss": 6.356776237487793,
+ "step": 269
+ },
+ {
+ "epoch": 3.7545851528384278,
+ "grad_norm": 0.012560366652905941,
+ "learning_rate": 0.0029829470089049815,
+ "loss": 6.357911109924316,
+ "step": 270
+ },
+ {
+ "epoch": 3.7685589519650655,
+ "grad_norm": 0.018435249105095863,
+ "learning_rate": 0.0029827449941757698,
+ "loss": 6.366821765899658,
+ "step": 271
+ },
+ {
+ "epoch": 3.7825327510917033,
+ "grad_norm": 0.029046855866909027,
+ "learning_rate": 0.002982541796910927,
+ "loss": 6.367595672607422,
+ "step": 272
+ },
+ {
+ "epoch": 3.7965065502183406,
+ "grad_norm": 0.043010637164115906,
+ "learning_rate": 0.002982337417274165,
+ "loss": 6.413588523864746,
+ "step": 273
+ },
+ {
+ "epoch": 3.810480349344978,
+ "grad_norm": 0.045470818877220154,
+ "learning_rate": 0.0029821318554301497,
+ "loss": 6.435015678405762,
+ "step": 274
+ },
+ {
+ "epoch": 3.8244541484716157,
+ "grad_norm": 0.02148726023733616,
+ "learning_rate": 0.002981925111544498,
+ "loss": 6.330894470214844,
+ "step": 275
+ },
+ {
+ "epoch": 3.8384279475982535,
+ "grad_norm": 0.026808688417077065,
+ "learning_rate": 0.00298171718578378,
+ "loss": 6.384839057922363,
+ "step": 276
+ },
+ {
+ "epoch": 3.8524017467248908,
+ "grad_norm": 0.018495213240385056,
+ "learning_rate": 0.002981508078315517,
+ "loss": 6.36277961730957,
+ "step": 277
+ },
+ {
+ "epoch": 3.8663755458515285,
+ "grad_norm": 0.025393597781658173,
+ "learning_rate": 0.0029812977893081834,
+ "loss": 6.3702239990234375,
+ "step": 278
+ },
+ {
+ "epoch": 3.880349344978166,
+ "grad_norm": 0.020856289193034172,
+ "learning_rate": 0.002981086318931206,
+ "loss": 6.354547500610352,
+ "step": 279
+ },
+ {
+ "epoch": 3.8943231441048036,
+ "grad_norm": 0.017827816307544708,
+ "learning_rate": 0.002980873667354962,
+ "loss": 6.339264869689941,
+ "step": 280
+ },
+ {
+ "epoch": 3.908296943231441,
+ "grad_norm": 0.02639714442193508,
+ "learning_rate": 0.002980659834750781,
+ "loss": 6.3254194259643555,
+ "step": 281
+ },
+ {
+ "epoch": 3.9222707423580787,
+ "grad_norm": 0.015238096937537193,
+ "learning_rate": 0.002980444821290944,
+ "loss": 6.365842819213867,
+ "step": 282
+ },
+ {
+ "epoch": 3.936244541484716,
+ "grad_norm": 0.012863729149103165,
+ "learning_rate": 0.002980228627148684,
+ "loss": 6.356776714324951,
+ "step": 283
+ },
+ {
+ "epoch": 3.950218340611354,
+ "grad_norm": 0.014739899896085262,
+ "learning_rate": 0.0029800112524981845,
+ "loss": 6.332481384277344,
+ "step": 284
+ },
+ {
+ "epoch": 3.964192139737991,
+ "grad_norm": 0.013582171872258186,
+ "learning_rate": 0.0029797926975145797,
+ "loss": 6.311968803405762,
+ "step": 285
+ },
+ {
+ "epoch": 3.978165938864629,
+ "grad_norm": 0.012704637832939625,
+ "learning_rate": 0.0029795729623739555,
+ "loss": 6.321654319763184,
+ "step": 286
+ },
+ {
+ "epoch": 3.992139737991266,
+ "grad_norm": 0.012550560757517815,
+ "learning_rate": 0.0029793520472533495,
+ "loss": 6.298409461975098,
+ "step": 287
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.011170165613293648,
+ "learning_rate": 0.002979129952330748,
+ "loss": 6.294830799102783,
+ "step": 288
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 6.326757907867432,
+ "eval_runtime": 51.6566,
+ "eval_samples_per_second": 47.274,
+ "eval_steps_per_second": 1.491,
+ "step": 288
+ },
+ {
+ "epoch": 4.013973799126638,
+ "grad_norm": 0.010947446338832378,
+ "learning_rate": 0.0029789066777850897,
+ "loss": 6.286581039428711,
+ "step": 289
+ },
+ {
+ "epoch": 4.0279475982532755,
+ "grad_norm": 0.010097995400428772,
+ "learning_rate": 0.002978682223796262,
+ "loss": 6.288301467895508,
+ "step": 290
+ },
+ {
+ "epoch": 4.041921397379912,
+ "grad_norm": 0.01224475260823965,
+ "learning_rate": 0.002978456590545104,
+ "loss": 6.264986991882324,
+ "step": 291
+ },
+ {
+ "epoch": 4.05589519650655,
+ "grad_norm": 0.009734537452459335,
+ "learning_rate": 0.0029782297782134045,
+ "loss": 6.304243087768555,
+ "step": 292
+ },
+ {
+ "epoch": 4.069868995633188,
+ "grad_norm": 0.01136956550180912,
+ "learning_rate": 0.0029780017869839013,
+ "loss": 6.244938373565674,
+ "step": 293
+ },
+ {
+ "epoch": 4.083842794759826,
+ "grad_norm": 0.00935293547809124,
+ "learning_rate": 0.002977772617040284,
+ "loss": 6.251852035522461,
+ "step": 294
+ },
+ {
+ "epoch": 4.097816593886463,
+ "grad_norm": 0.011118784546852112,
+ "learning_rate": 0.0029775422685671894,
+ "loss": 6.256633758544922,
+ "step": 295
+ },
+ {
+ "epoch": 4.1117903930131,
+ "grad_norm": 0.010837923735380173,
+ "learning_rate": 0.0029773107417502055,
+ "loss": 6.263411045074463,
+ "step": 296
+ },
+ {
+ "epoch": 4.125764192139738,
+ "grad_norm": 0.011920305900275707,
+ "learning_rate": 0.0029770780367758703,
+ "loss": 6.239311218261719,
+ "step": 297
+ },
+ {
+ "epoch": 4.139737991266376,
+ "grad_norm": 0.014117483049631119,
+ "learning_rate": 0.0029768441538316685,
+ "loss": 6.253314018249512,
+ "step": 298
+ },
+ {
+ "epoch": 4.153711790393013,
+ "grad_norm": 0.022502148523926735,
+ "learning_rate": 0.002976609093106036,
+ "loss": 6.246156692504883,
+ "step": 299
+ },
+ {
+ "epoch": 4.1676855895196505,
+ "grad_norm": 0.03348720073699951,
+ "learning_rate": 0.002976372854788357,
+ "loss": 6.255914688110352,
+ "step": 300
+ },
+ {
+ "epoch": 4.181659388646288,
+ "grad_norm": 0.03427121415734291,
+ "learning_rate": 0.002976135439068965,
+ "loss": 6.289946556091309,
+ "step": 301
+ },
+ {
+ "epoch": 4.195633187772926,
+ "grad_norm": 0.02407355234026909,
+ "learning_rate": 0.002975896846139141,
+ "loss": 6.288534164428711,
+ "step": 302
+ },
+ {
+ "epoch": 4.209606986899563,
+ "grad_norm": 0.017861567437648773,
+ "learning_rate": 0.0029756570761911147,
+ "loss": 6.258049964904785,
+ "step": 303
+ },
+ {
+ "epoch": 4.223580786026201,
+ "grad_norm": 0.019980227574706078,
+ "learning_rate": 0.0029754161294180646,
+ "loss": 6.250895023345947,
+ "step": 304
+ },
+ {
+ "epoch": 4.2375545851528384,
+ "grad_norm": 0.01443127728998661,
+ "learning_rate": 0.0029751740060141177,
+ "loss": 6.212200164794922,
+ "step": 305
+ },
+ {
+ "epoch": 4.251528384279476,
+ "grad_norm": 0.014975055120885372,
+ "learning_rate": 0.0029749307061743475,
+ "loss": 6.230568885803223,
+ "step": 306
+ },
+ {
+ "epoch": 4.265502183406113,
+ "grad_norm": 0.01408296450972557,
+ "learning_rate": 0.0029746862300947773,
+ "loss": 6.243950366973877,
+ "step": 307
+ },
+ {
+ "epoch": 4.279475982532751,
+ "grad_norm": 0.011649327352643013,
+ "learning_rate": 0.0029744405779723766,
+ "loss": 6.197744846343994,
+ "step": 308
+ },
+ {
+ "epoch": 4.293449781659389,
+ "grad_norm": 0.013392188586294651,
+ "learning_rate": 0.002974193750005063,
+ "loss": 6.207277297973633,
+ "step": 309
+ },
+ {
+ "epoch": 4.307423580786026,
+ "grad_norm": 0.010902997106313705,
+ "learning_rate": 0.002973945746391701,
+ "loss": 6.166373252868652,
+ "step": 310
+ },
+ {
+ "epoch": 4.321397379912664,
+ "grad_norm": 0.012915392406284809,
+ "learning_rate": 0.0029736965673321032,
+ "loss": 6.186316013336182,
+ "step": 311
+ },
+ {
+ "epoch": 4.335371179039301,
+ "grad_norm": 0.010772830806672573,
+ "learning_rate": 0.0029734462130270273,
+ "loss": 6.172578811645508,
+ "step": 312
+ },
+ {
+ "epoch": 4.349344978165939,
+ "grad_norm": 0.010274911299347878,
+ "learning_rate": 0.002973194683678181,
+ "loss": 6.206851005554199,
+ "step": 313
+ },
+ {
+ "epoch": 4.3633187772925766,
+ "grad_norm": 0.013210494071245193,
+ "learning_rate": 0.0029729419794882164,
+ "loss": 6.21115779876709,
+ "step": 314
+ },
+ {
+ "epoch": 4.377292576419214,
+ "grad_norm": 0.013372313231229782,
+ "learning_rate": 0.0029726881006607326,
+ "loss": 6.18442440032959,
+ "step": 315
+ },
+ {
+ "epoch": 4.391266375545851,
+ "grad_norm": 0.008105908520519733,
+ "learning_rate": 0.0029724330474002743,
+ "loss": 6.204354286193848,
+ "step": 316
+ },
+ {
+ "epoch": 4.405240174672489,
+ "grad_norm": 0.014167108573019505,
+ "learning_rate": 0.0029721768199123346,
+ "loss": 6.180662155151367,
+ "step": 317
+ },
+ {
+ "epoch": 4.419213973799127,
+ "grad_norm": 0.008339803665876389,
+ "learning_rate": 0.002971919418403351,
+ "loss": 6.150079727172852,
+ "step": 318
+ },
+ {
+ "epoch": 4.4331877729257645,
+ "grad_norm": 0.015939386561512947,
+ "learning_rate": 0.0029716608430807073,
+ "loss": 6.175203323364258,
+ "step": 319
+ },
+ {
+ "epoch": 4.447161572052401,
+ "grad_norm": 0.010777109302580357,
+ "learning_rate": 0.0029714010941527327,
+ "loss": 6.166106224060059,
+ "step": 320
+ },
+ {
+ "epoch": 4.461135371179039,
+ "grad_norm": 0.011274735443294048,
+ "learning_rate": 0.002971140171828702,
+ "loss": 6.163111686706543,
+ "step": 321
+ },
+ {
+ "epoch": 4.475109170305677,
+ "grad_norm": 0.013626808300614357,
+ "learning_rate": 0.002970878076318836,
+ "loss": 6.140850067138672,
+ "step": 322
+ },
+ {
+ "epoch": 4.489082969432315,
+ "grad_norm": 0.016907762736082077,
+ "learning_rate": 0.0029706148078343,
+ "loss": 6.192012786865234,
+ "step": 323
+ },
+ {
+ "epoch": 4.503056768558952,
+ "grad_norm": 0.018817441537976265,
+ "learning_rate": 0.002970350366587205,
+ "loss": 6.162238121032715,
+ "step": 324
+ },
+ {
+ "epoch": 4.517030567685589,
+ "grad_norm": 0.02829768694937229,
+ "learning_rate": 0.0029700847527906068,
+ "loss": 6.159066200256348,
+ "step": 325
+ },
+ {
+ "epoch": 4.531004366812227,
+ "grad_norm": 0.02034052088856697,
+ "learning_rate": 0.002969817966658505,
+ "loss": 6.164158344268799,
+ "step": 326
+ },
+ {
+ "epoch": 4.544978165938865,
+ "grad_norm": 0.01816369965672493,
+ "learning_rate": 0.0029695500084058447,
+ "loss": 6.132316589355469,
+ "step": 327
+ },
+ {
+ "epoch": 4.558951965065502,
+ "grad_norm": 0.013685261830687523,
+ "learning_rate": 0.0029692808782485146,
+ "loss": 6.152181625366211,
+ "step": 328
+ },
+ {
+ "epoch": 4.5729257641921395,
+ "grad_norm": 0.009791149757802486,
+ "learning_rate": 0.0029690105764033485,
+ "loss": 6.138544082641602,
+ "step": 329
+ },
+ {
+ "epoch": 4.586899563318777,
+ "grad_norm": 0.014827069826424122,
+ "learning_rate": 0.002968739103088124,
+ "loss": 6.135766983032227,
+ "step": 330
+ },
+ {
+ "epoch": 4.600873362445415,
+ "grad_norm": 0.01391249243170023,
+ "learning_rate": 0.0029684664585215618,
+ "loss": 6.146172523498535,
+ "step": 331
+ },
+ {
+ "epoch": 4.614847161572053,
+ "grad_norm": 0.012729899026453495,
+ "learning_rate": 0.002968192642923326,
+ "loss": 6.122613430023193,
+ "step": 332
+ },
+ {
+ "epoch": 4.62882096069869,
+ "grad_norm": 0.014175458811223507,
+ "learning_rate": 0.002967917656514026,
+ "loss": 6.12574577331543,
+ "step": 333
+ },
+ {
+ "epoch": 4.642794759825327,
+ "grad_norm": 0.010958840139210224,
+ "learning_rate": 0.002967641499515213,
+ "loss": 6.111088752746582,
+ "step": 334
+ },
+ {
+ "epoch": 4.656768558951965,
+ "grad_norm": 0.015380816534161568,
+ "learning_rate": 0.0029673641721493817,
+ "loss": 6.100615501403809,
+ "step": 335
+ },
+ {
+ "epoch": 4.670742358078603,
+ "grad_norm": 0.025339096784591675,
+ "learning_rate": 0.00296708567463997,
+ "loss": 6.139659881591797,
+ "step": 336
+ },
+ {
+ "epoch": 4.68471615720524,
+ "grad_norm": 0.03695995733141899,
+ "learning_rate": 0.002966806007211358,
+ "loss": 6.160996437072754,
+ "step": 337
+ },
+ {
+ "epoch": 4.698689956331878,
+ "grad_norm": 0.0408046580851078,
+ "learning_rate": 0.0029665251700888686,
+ "loss": 6.173980712890625,
+ "step": 338
+ },
+ {
+ "epoch": 4.712663755458515,
+ "grad_norm": 0.025022024288773537,
+ "learning_rate": 0.0029662431634987674,
+ "loss": 6.145794868469238,
+ "step": 339
+ },
+ {
+ "epoch": 4.726637554585153,
+ "grad_norm": 0.027254154905676842,
+ "learning_rate": 0.002965959987668262,
+ "loss": 6.16928243637085,
+ "step": 340
+ },
+ {
+ "epoch": 4.74061135371179,
+ "grad_norm": 0.06191903352737427,
+ "learning_rate": 0.0029656756428255017,
+ "loss": 6.138525009155273,
+ "step": 341
+ },
+ {
+ "epoch": 4.754585152838428,
+ "grad_norm": 0.0348956398665905,
+ "learning_rate": 0.002965390129199579,
+ "loss": 6.185477256774902,
+ "step": 342
+ },
+ {
+ "epoch": 4.7685589519650655,
+ "grad_norm": 0.020859677344560623,
+ "learning_rate": 0.002965103447020526,
+ "loss": 6.148846626281738,
+ "step": 343
+ },
+ {
+ "epoch": 4.782532751091703,
+ "grad_norm": 0.022810814902186394,
+ "learning_rate": 0.0029648155965193174,
+ "loss": 6.1701178550720215,
+ "step": 344
+ },
+ {
+ "epoch": 4.796506550218341,
+ "grad_norm": 0.017981817945837975,
+ "learning_rate": 0.00296452657792787,
+ "loss": 6.1323771476745605,
+ "step": 345
+ },
+ {
+ "epoch": 4.810480349344978,
+ "grad_norm": 0.014905866235494614,
+ "learning_rate": 0.00296423639147904,
+ "loss": 6.123426914215088,
+ "step": 346
+ },
+ {
+ "epoch": 4.824454148471616,
+ "grad_norm": 0.01463468000292778,
+ "learning_rate": 0.002963945037406626,
+ "loss": 6.146122932434082,
+ "step": 347
+ },
+ {
+ "epoch": 4.8384279475982535,
+ "grad_norm": 0.021616583690047264,
+ "learning_rate": 0.002963652515945366,
+ "loss": 6.093459606170654,
+ "step": 348
+ },
+ {
+ "epoch": 4.85240174672489,
+ "grad_norm": 0.01550968922674656,
+ "learning_rate": 0.00296335882733094,
+ "loss": 6.114929676055908,
+ "step": 349
+ },
+ {
+ "epoch": 4.866375545851528,
+ "grad_norm": 0.014597428031265736,
+ "learning_rate": 0.0029630639717999675,
+ "loss": 6.113945960998535,
+ "step": 350
+ },
+ {
+ "epoch": 4.880349344978166,
+ "grad_norm": 0.018249280750751495,
+ "learning_rate": 0.002962767949590008,
+ "loss": 6.130507469177246,
+ "step": 351
+ },
+ {
+ "epoch": 4.894323144104804,
+ "grad_norm": 0.014160148799419403,
+ "learning_rate": 0.0029624707609395614,
+ "loss": 6.086946487426758,
+ "step": 352
+ },
+ {
+ "epoch": 4.908296943231441,
+ "grad_norm": 0.015563832595944405,
+ "learning_rate": 0.002962172406088067,
+ "loss": 6.083332538604736,
+ "step": 353
+ },
+ {
+ "epoch": 4.922270742358078,
+ "grad_norm": 0.012240015901625156,
+ "learning_rate": 0.002961872885275904,
+ "loss": 6.0822906494140625,
+ "step": 354
+ },
+ {
+ "epoch": 4.936244541484716,
+ "grad_norm": 0.0141651201993227,
+ "learning_rate": 0.002961572198744391,
+ "loss": 6.049562454223633,
+ "step": 355
+ },
+ {
+ "epoch": 4.950218340611354,
+ "grad_norm": 0.010004645213484764,
+ "learning_rate": 0.0029612703467357866,
+ "loss": 6.067063808441162,
+ "step": 356
+ },
+ {
+ "epoch": 4.964192139737992,
+ "grad_norm": 0.011790835298597813,
+ "learning_rate": 0.0029609673294932856,
+ "loss": 6.062193870544434,
+ "step": 357
+ },
+ {
+ "epoch": 4.978165938864628,
+ "grad_norm": 0.011088060215115547,
+ "learning_rate": 0.002960663147261025,
+ "loss": 6.070950031280518,
+ "step": 358
+ },
+ {
+ "epoch": 4.992139737991266,
+ "grad_norm": 0.010387620888650417,
+ "learning_rate": 0.002960357800284078,
+ "loss": 6.032423973083496,
+ "step": 359
+ },
+ {
+ "epoch": 5.0,
+ "grad_norm": 0.012074458412826061,
+ "learning_rate": 0.002960051288808458,
+ "loss": 6.066788673400879,
+ "step": 360
+ },
+ {
+ "epoch": 5.0,
+ "eval_loss": 6.097111701965332,
+ "eval_runtime": 51.8742,
+ "eval_samples_per_second": 47.075,
+ "eval_steps_per_second": 1.484,
+ "step": 360
+ },
+ {
+ "epoch": 5.013973799126638,
+ "grad_norm": 0.013429952785372734,
+ "learning_rate": 0.0029597436130811153,
+ "loss": 6.022172451019287,
+ "step": 361
+ },
+ {
+ "epoch": 5.0279475982532755,
+ "grad_norm": 0.01764705777168274,
+ "learning_rate": 0.0029594347733499387,
+ "loss": 6.029745101928711,
+ "step": 362
+ },
+ {
+ "epoch": 5.041921397379912,
+ "grad_norm": 0.02403099089860916,
+ "learning_rate": 0.002959124769863755,
+ "loss": 6.04707145690918,
+ "step": 363
+ },
+ {
+ "epoch": 5.05589519650655,
+ "grad_norm": 0.029030559584498405,
+ "learning_rate": 0.002958813602872328,
+ "loss": 6.05479097366333,
+ "step": 364
+ },
+ {
+ "epoch": 5.069868995633188,
+ "grad_norm": 0.029980644583702087,
+ "learning_rate": 0.0029585012726263595,
+ "loss": 6.068657875061035,
+ "step": 365
+ },
+ {
+ "epoch": 5.083842794759826,
+ "grad_norm": 0.021424276754260063,
+ "learning_rate": 0.0029581877793774885,
+ "loss": 6.062008857727051,
+ "step": 366
+ },
+ {
+ "epoch": 5.097816593886463,
+ "grad_norm": 0.018076395615935326,
+ "learning_rate": 0.002957873123378291,
+ "loss": 6.023279190063477,
+ "step": 367
+ },
+ {
+ "epoch": 5.1117903930131,
+ "grad_norm": 0.017688792198896408,
+ "learning_rate": 0.002957557304882279,
+ "loss": 6.014641284942627,
+ "step": 368
+ },
+ {
+ "epoch": 5.125764192139738,
+ "grad_norm": 0.02313626930117607,
+ "learning_rate": 0.0029572403241439026,
+ "loss": 6.026420593261719,
+ "step": 369
+ },
+ {
+ "epoch": 5.139737991266376,
+ "grad_norm": 0.04045062139630318,
+ "learning_rate": 0.0029569221814185467,
+ "loss": 6.007199764251709,
+ "step": 370
+ },
+ {
+ "epoch": 5.153711790393013,
+ "grad_norm": 0.015792209655046463,
+ "learning_rate": 0.0029566028769625334,
+ "loss": 6.017209053039551,
+ "step": 371
+ },
+ {
+ "epoch": 5.1676855895196505,
+ "grad_norm": 0.01990479789674282,
+ "learning_rate": 0.002956282411033121,
+ "loss": 5.982841491699219,
+ "step": 372
+ },
+ {
+ "epoch": 5.181659388646288,
+ "grad_norm": 0.01191043108701706,
+ "learning_rate": 0.002955960783888503,
+ "loss": 6.0035929679870605,
+ "step": 373
+ },
+ {
+ "epoch": 5.195633187772926,
+ "grad_norm": 0.01732831820845604,
+ "learning_rate": 0.002955637995787807,
+ "loss": 6.025006294250488,
+ "step": 374
+ },
+ {
+ "epoch": 5.209606986899563,
+ "grad_norm": 0.032948266714811325,
+ "learning_rate": 0.0029553140469910995,
+ "loss": 6.007736682891846,
+ "step": 375
+ },
+ {
+ "epoch": 5.223580786026201,
+ "grad_norm": 0.011192727833986282,
+ "learning_rate": 0.00295498893775938,
+ "loss": 6.0022687911987305,
+ "step": 376
+ },
+ {
+ "epoch": 5.2375545851528384,
+ "grad_norm": 0.014646840281784534,
+ "learning_rate": 0.0029546626683545824,
+ "loss": 5.976479530334473,
+ "step": 377
+ },
+ {
+ "epoch": 5.251528384279476,
+ "grad_norm": 0.01589796133339405,
+ "learning_rate": 0.002954335239039576,
+ "loss": 5.980949878692627,
+ "step": 378
+ },
+ {
+ "epoch": 5.265502183406113,
+ "grad_norm": 0.015129734762012959,
+ "learning_rate": 0.0029540066500781657,
+ "loss": 6.0073394775390625,
+ "step": 379
+ },
+ {
+ "epoch": 5.279475982532751,
+ "grad_norm": 0.011140518821775913,
+ "learning_rate": 0.002953676901735088,
+ "loss": 5.976587295532227,
+ "step": 380
+ },
+ {
+ "epoch": 5.293449781659389,
+ "grad_norm": 0.0109480582177639,
+ "learning_rate": 0.0029533459942760174,
+ "loss": 6.001427173614502,
+ "step": 381
+ },
+ {
+ "epoch": 5.307423580786026,
+ "grad_norm": 0.013346790336072445,
+ "learning_rate": 0.0029530139279675583,
+ "loss": 5.96919059753418,
+ "step": 382
+ },
+ {
+ "epoch": 5.321397379912664,
+ "grad_norm": 0.012474581599235535,
+ "learning_rate": 0.002952680703077251,
+ "loss": 5.970426559448242,
+ "step": 383
+ },
+ {
+ "epoch": 5.335371179039301,
+ "grad_norm": 0.01140566635876894,
+ "learning_rate": 0.002952346319873569,
+ "loss": 5.989171504974365,
+ "step": 384
+ },
+ {
+ "epoch": 5.349344978165939,
+ "grad_norm": 0.0137029392644763,
+ "learning_rate": 0.0029520107786259187,
+ "loss": 6.010115623474121,
+ "step": 385
+ },
+ {
+ "epoch": 5.3633187772925766,
+ "grad_norm": 0.015517761930823326,
+ "learning_rate": 0.002951674079604639,
+ "loss": 6.003649711608887,
+ "step": 386
+ },
+ {
+ "epoch": 5.377292576419214,
+ "grad_norm": 0.02173629403114319,
+ "learning_rate": 0.0029513362230810035,
+ "loss": 5.967074871063232,
+ "step": 387
+ },
+ {
+ "epoch": 5.391266375545851,
+ "grad_norm": 0.02683914266526699,
+ "learning_rate": 0.002950997209327216,
+ "loss": 5.994434833526611,
+ "step": 388
+ },
+ {
+ "epoch": 5.405240174672489,
+ "grad_norm": 0.025865858420729637,
+ "learning_rate": 0.002950657038616415,
+ "loss": 6.010523796081543,
+ "step": 389
+ },
+ {
+ "epoch": 5.419213973799127,
+ "grad_norm": 0.02544897235929966,
+ "learning_rate": 0.002950315711222668,
+ "loss": 5.994078159332275,
+ "step": 390
+ },
+ {
+ "epoch": 5.4331877729257645,
+ "grad_norm": 0.028426019474864006,
+ "learning_rate": 0.0029499732274209777,
+ "loss": 6.020925521850586,
+ "step": 391
+ },
+ {
+ "epoch": 5.447161572052401,
+ "grad_norm": 0.026174750179052353,
+ "learning_rate": 0.0029496295874872768,
+ "loss": 5.985616683959961,
+ "step": 392
+ },
+ {
+ "epoch": 5.461135371179039,
+ "grad_norm": 0.02802138403058052,
+ "learning_rate": 0.00294928479169843,
+ "loss": 5.973659038543701,
+ "step": 393
+ },
+ {
+ "epoch": 5.475109170305677,
+ "grad_norm": 0.019348938018083572,
+ "learning_rate": 0.002948938840332233,
+ "loss": 5.990137577056885,
+ "step": 394
+ },
+ {
+ "epoch": 5.489082969432315,
+ "grad_norm": 0.016294041648507118,
+ "learning_rate": 0.0029485917336674115,
+ "loss": 5.962082862854004,
+ "step": 395
+ },
+ {
+ "epoch": 5.503056768558952,
+ "grad_norm": 0.016167785972356796,
+ "learning_rate": 0.002948243471983625,
+ "loss": 5.9594573974609375,
+ "step": 396
+ },
+ {
+ "epoch": 5.517030567685589,
+ "grad_norm": 0.014119232073426247,
+ "learning_rate": 0.00294789405556146,
+ "loss": 5.894467353820801,
+ "step": 397
+ },
+ {
+ "epoch": 5.531004366812227,
+ "grad_norm": 0.013805347494781017,
+ "learning_rate": 0.002947543484682436,
+ "loss": 5.964014530181885,
+ "step": 398
+ },
+ {
+ "epoch": 5.544978165938865,
+ "grad_norm": 0.014496497809886932,
+ "learning_rate": 0.002947191759629001,
+ "loss": 5.989920616149902,
+ "step": 399
+ },
+ {
+ "epoch": 5.558951965065502,
+ "grad_norm": 0.014813800342381,
+ "learning_rate": 0.0029468388806845333,
+ "loss": 5.9669413566589355,
+ "step": 400
+ },
+ {
+ "epoch": 5.5729257641921395,
+ "grad_norm": 0.01153622381389141,
+ "learning_rate": 0.0029464848481333423,
+ "loss": 5.918018341064453,
+ "step": 401
+ },
+ {
+ "epoch": 5.586899563318777,
+ "grad_norm": 0.010525977239012718,
+ "learning_rate": 0.0029461296622606647,
+ "loss": 5.892477989196777,
+ "step": 402
+ },
+ {
+ "epoch": 5.600873362445415,
+ "grad_norm": 0.011410463601350784,
+ "learning_rate": 0.002945773323352667,
+ "loss": 5.925249099731445,
+ "step": 403
+ },
+ {
+ "epoch": 5.614847161572053,
+ "grad_norm": 0.010949424467980862,
+ "learning_rate": 0.0029454158316964464,
+ "loss": 5.90645694732666,
+ "step": 404
+ },
+ {
+ "epoch": 5.62882096069869,
+ "grad_norm": 0.00962804164737463,
+ "learning_rate": 0.002945057187580026,
+ "loss": 5.917054176330566,
+ "step": 405
+ },
+ {
+ "epoch": 5.642794759825327,
+ "grad_norm": 0.011517546139657497,
+ "learning_rate": 0.0029446973912923597,
+ "loss": 5.919445514678955,
+ "step": 406
+ },
+ {
+ "epoch": 5.656768558951965,
+ "grad_norm": 0.01567666605114937,
+ "learning_rate": 0.0029443364431233285,
+ "loss": 5.884111404418945,
+ "step": 407
+ },
+ {
+ "epoch": 5.670742358078603,
+ "grad_norm": 0.016781140118837357,
+ "learning_rate": 0.002943974343363742,
+ "loss": 5.9252119064331055,
+ "step": 408
+ },
+ {
+ "epoch": 5.68471615720524,
+ "grad_norm": 0.02244546078145504,
+ "learning_rate": 0.0029436110923053375,
+ "loss": 5.933143615722656,
+ "step": 409
+ },
+ {
+ "epoch": 5.698689956331878,
+ "grad_norm": 0.028186792507767677,
+ "learning_rate": 0.00294324669024078,
+ "loss": 5.9079132080078125,
+ "step": 410
+ },
+ {
+ "epoch": 5.712663755458515,
+ "grad_norm": 0.028585223481059074,
+ "learning_rate": 0.002942881137463661,
+ "loss": 5.925262451171875,
+ "step": 411
+ },
+ {
+ "epoch": 5.726637554585153,
+ "grad_norm": 0.02781485952436924,
+ "learning_rate": 0.0029425144342685003,
+ "loss": 5.963315010070801,
+ "step": 412
+ },
+ {
+ "epoch": 5.74061135371179,
+ "grad_norm": 0.021486928686499596,
+ "learning_rate": 0.002942146580950744,
+ "loss": 5.930115699768066,
+ "step": 413
+ },
+ {
+ "epoch": 5.754585152838428,
+ "grad_norm": 0.016890818253159523,
+ "learning_rate": 0.002941777577806765,
+ "loss": 5.921439170837402,
+ "step": 414
+ },
+ {
+ "epoch": 5.7685589519650655,
+ "grad_norm": 0.0169126745313406,
+ "learning_rate": 0.0029414074251338618,
+ "loss": 5.928220272064209,
+ "step": 415
+ },
+ {
+ "epoch": 5.782532751091703,
+ "grad_norm": 0.014659176580607891,
+ "learning_rate": 0.002941036123230261,
+ "loss": 5.934546947479248,
+ "step": 416
+ },
+ {
+ "epoch": 5.796506550218341,
+ "grad_norm": 0.015337883494794369,
+ "learning_rate": 0.002940663672395112,
+ "loss": 5.890145301818848,
+ "step": 417
+ },
+ {
+ "epoch": 5.810480349344978,
+ "grad_norm": 0.01254988368600607,
+ "learning_rate": 0.002940290072928493,
+ "loss": 5.875492095947266,
+ "step": 418
+ },
+ {
+ "epoch": 5.824454148471616,
+ "grad_norm": 0.014658362604677677,
+ "learning_rate": 0.0029399153251314066,
+ "loss": 5.901272773742676,
+ "step": 419
+ },
+ {
+ "epoch": 5.8384279475982535,
+ "grad_norm": 0.017342250794172287,
+ "learning_rate": 0.002939539429305779,
+ "loss": 5.875853538513184,
+ "step": 420
+ },
+ {
+ "epoch": 5.85240174672489,
+ "grad_norm": 0.016825608909130096,
+ "learning_rate": 0.0029391623857544643,
+ "loss": 5.941972732543945,
+ "step": 421
+ },
+ {
+ "epoch": 5.866375545851528,
+ "grad_norm": 0.016716884449124336,
+ "learning_rate": 0.002938784194781238,
+ "loss": 5.912367343902588,
+ "step": 422
+ },
+ {
+ "epoch": 5.880349344978166,
+ "grad_norm": 0.021707067266106606,
+ "learning_rate": 0.002938404856690803,
+ "loss": 5.895319938659668,
+ "step": 423
+ },
+ {
+ "epoch": 5.894323144104804,
+ "grad_norm": 0.030519135296344757,
+ "learning_rate": 0.002938024371788785,
+ "loss": 5.897165298461914,
+ "step": 424
+ },
+ {
+ "epoch": 5.908296943231441,
+ "grad_norm": 0.02572188340127468,
+ "learning_rate": 0.002937642740381733,
+ "loss": 5.879164695739746,
+ "step": 425
+ },
+ {
+ "epoch": 5.922270742358078,
+ "grad_norm": 0.016286717727780342,
+ "learning_rate": 0.002937259962777122,
+ "loss": 5.841421127319336,
+ "step": 426
+ },
+ {
+ "epoch": 5.936244541484716,
+ "grad_norm": 0.020656082779169083,
+ "learning_rate": 0.002936876039283347,
+ "loss": 5.9222259521484375,
+ "step": 427
+ },
+ {
+ "epoch": 5.950218340611354,
+ "grad_norm": 0.015350459143519402,
+ "learning_rate": 0.0029364909702097298,
+ "loss": 5.885847091674805,
+ "step": 428
+ },
+ {
+ "epoch": 5.964192139737992,
+ "grad_norm": 0.01708570495247841,
+ "learning_rate": 0.0029361047558665136,
+ "loss": 5.872044563293457,
+ "step": 429
+ },
+ {
+ "epoch": 5.978165938864628,
+ "grad_norm": 0.013537376187741756,
+ "learning_rate": 0.0029357173965648634,
+ "loss": 5.866434097290039,
+ "step": 430
+ },
+ {
+ "epoch": 5.992139737991266,
+ "grad_norm": 0.013208637945353985,
+ "learning_rate": 0.002935328892616868,
+ "loss": 5.8624491691589355,
+ "step": 431
+ },
+ {
+ "epoch": 6.0,
+ "grad_norm": 0.012249554507434368,
+ "learning_rate": 0.002934939244335538,
+ "loss": 5.833232879638672,
+ "step": 432
+ },
+ {
+ "epoch": 6.0,
+ "eval_loss": 5.90985107421875,
+ "eval_runtime": 51.0696,
+ "eval_samples_per_second": 47.817,
+ "eval_steps_per_second": 1.508,
+ "step": 432
+ },
+ {
+ "epoch": 6.013973799126638,
+ "grad_norm": 0.01382362935692072,
+ "learning_rate": 0.0029345484520348062,
+ "loss": 5.835695266723633,
+ "step": 433
+ },
+ {
+ "epoch": 6.0279475982532755,
+ "grad_norm": 0.012947693467140198,
+ "learning_rate": 0.0029341565160295262,
+ "loss": 5.826190948486328,
+ "step": 434
+ },
+ {
+ "epoch": 6.041921397379912,
+ "grad_norm": 0.013653872534632683,
+ "learning_rate": 0.002933763436635475,
+ "loss": 5.831789016723633,
+ "step": 435
+ },
+ {
+ "epoch": 6.05589519650655,
+ "grad_norm": 0.027596980333328247,
+ "learning_rate": 0.0029333692141693484,
+ "loss": 5.839963436126709,
+ "step": 436
+ },
+ {
+ "epoch": 6.069868995633188,
+ "grad_norm": 0.01637990213930607,
+ "learning_rate": 0.0029329738489487648,
+ "loss": 5.805248260498047,
+ "step": 437
+ },
+ {
+ "epoch": 6.083842794759826,
+ "grad_norm": 0.02169182151556015,
+ "learning_rate": 0.0029325773412922627,
+ "loss": 5.838248252868652,
+ "step": 438
+ },
+ {
+ "epoch": 6.097816593886463,
+ "grad_norm": 0.03518201410770416,
+ "learning_rate": 0.0029321796915193015,
+ "loss": 5.834336280822754,
+ "step": 439
+ },
+ {
+ "epoch": 6.1117903930131,
+ "grad_norm": 0.04983877018094063,
+ "learning_rate": 0.0029317808999502603,
+ "loss": 5.879856109619141,
+ "step": 440
+ },
+ {
+ "epoch": 6.125764192139738,
+ "grad_norm": 0.05085664615035057,
+ "learning_rate": 0.0029313809669064377,
+ "loss": 5.946531295776367,
+ "step": 441
+ },
+ {
+ "epoch": 6.139737991266376,
+ "grad_norm": 0.03314858302474022,
+ "learning_rate": 0.0029309798927100532,
+ "loss": 5.909746170043945,
+ "step": 442
+ },
+ {
+ "epoch": 6.153711790393013,
+ "grad_norm": 0.034118980169296265,
+ "learning_rate": 0.0029305776776842453,
+ "loss": 5.903947830200195,
+ "step": 443
+ },
+ {
+ "epoch": 6.1676855895196505,
+ "grad_norm": 0.03401247411966324,
+ "learning_rate": 0.0029301743221530705,
+ "loss": 5.936162948608398,
+ "step": 444
+ },
+ {
+ "epoch": 6.181659388646288,
+ "grad_norm": 0.043827857822179794,
+ "learning_rate": 0.0029297698264415047,
+ "loss": 5.953949451446533,
+ "step": 445
+ },
+ {
+ "epoch": 6.195633187772926,
+ "grad_norm": 0.06281330436468124,
+ "learning_rate": 0.002929364190875444,
+ "loss": 5.988225936889648,
+ "step": 446
+ },
+ {
+ "epoch": 6.209606986899563,
+ "grad_norm": 0.07216024398803711,
+ "learning_rate": 0.0029289574157817014,
+ "loss": 6.187741279602051,
+ "step": 447
+ },
+ {
+ "epoch": 6.223580786026201,
+ "grad_norm": 0.055205658078193665,
+ "learning_rate": 0.002928549501488008,
+ "loss": 6.177831649780273,
+ "step": 448
+ },
+ {
+ "epoch": 6.2375545851528384,
+ "grad_norm": 0.04834286496043205,
+ "learning_rate": 0.0029281404483230133,
+ "loss": 6.209604263305664,
+ "step": 449
+ },
+ {
+ "epoch": 6.251528384279476,
+ "grad_norm": 0.0447777584195137,
+ "learning_rate": 0.0029277302566162837,
+ "loss": 6.146507263183594,
+ "step": 450
+ },
+ {
+ "epoch": 6.265502183406113,
+ "grad_norm": 0.05729430541396141,
+ "learning_rate": 0.0029273189266983035,
+ "loss": 6.280686378479004,
+ "step": 451
+ },
+ {
+ "epoch": 6.279475982532751,
+ "grad_norm": 0.0410807766020298,
+ "learning_rate": 0.002926906458900473,
+ "loss": 6.246824264526367,
+ "step": 452
+ },
+ {
+ "epoch": 6.293449781659389,
+ "grad_norm": 0.044607777148485184,
+ "learning_rate": 0.0029264928535551104,
+ "loss": 6.19956111907959,
+ "step": 453
+ },
+ {
+ "epoch": 6.307423580786026,
+ "grad_norm": 0.05141043663024902,
+ "learning_rate": 0.0029260781109954515,
+ "loss": 6.118578910827637,
+ "step": 454
+ },
+ {
+ "epoch": 6.321397379912664,
+ "grad_norm": 0.029248764738440514,
+ "learning_rate": 0.0029256622315556448,
+ "loss": 6.155360221862793,
+ "step": 455
+ },
+ {
+ "epoch": 6.335371179039301,
+ "grad_norm": 0.028403814882040024,
+ "learning_rate": 0.0029252452155707582,
+ "loss": 6.097352027893066,
+ "step": 456
+ },
+ {
+ "epoch": 6.349344978165939,
+ "grad_norm": 0.025872666388750076,
+ "learning_rate": 0.002924827063376773,
+ "loss": 6.072485446929932,
+ "step": 457
+ },
+ {
+ "epoch": 6.3633187772925766,
+ "grad_norm": 0.018447555601596832,
+ "learning_rate": 0.0029244077753105874,
+ "loss": 6.077045440673828,
+ "step": 458
+ },
+ {
+ "epoch": 6.377292576419214,
+ "grad_norm": 0.023166058585047722,
+ "learning_rate": 0.0029239873517100147,
+ "loss": 6.040050983428955,
+ "step": 459
+ },
+ {
+ "epoch": 6.391266375545851,
+ "grad_norm": 0.019579261541366577,
+ "learning_rate": 0.002923565792913782,
+ "loss": 6.039424896240234,
+ "step": 460
+ },
+ {
+ "epoch": 6.405240174672489,
+ "grad_norm": 0.018369073048233986,
+ "learning_rate": 0.002923143099261531,
+ "loss": 6.021180152893066,
+ "step": 461
+ },
+ {
+ "epoch": 6.419213973799127,
+ "grad_norm": 0.014188461937010288,
+ "learning_rate": 0.0029227192710938195,
+ "loss": 5.9668049812316895,
+ "step": 462
+ },
+ {
+ "epoch": 6.4331877729257645,
+ "grad_norm": 0.019567398354411125,
+ "learning_rate": 0.0029222943087521182,
+ "loss": 5.989189147949219,
+ "step": 463
+ },
+ {
+ "epoch": 6.447161572052401,
+ "grad_norm": 0.014190621674060822,
+ "learning_rate": 0.0029218682125788102,
+ "loss": 5.942726135253906,
+ "step": 464
+ },
+ {
+ "epoch": 6.461135371179039,
+ "grad_norm": 0.01375526562333107,
+ "learning_rate": 0.002921440982917195,
+ "loss": 5.942777633666992,
+ "step": 465
+ },
+ {
+ "epoch": 6.475109170305677,
+ "grad_norm": 0.014203300699591637,
+ "learning_rate": 0.0029210126201114836,
+ "loss": 5.927928924560547,
+ "step": 466
+ },
+ {
+ "epoch": 6.489082969432315,
+ "grad_norm": 0.014231069944798946,
+ "learning_rate": 0.0029205831245067994,
+ "loss": 5.923510551452637,
+ "step": 467
+ },
+ {
+ "epoch": 6.503056768558952,
+ "grad_norm": 0.012771515175700188,
+ "learning_rate": 0.0029201524964491796,
+ "loss": 5.908803939819336,
+ "step": 468
+ },
+ {
+ "epoch": 6.517030567685589,
+ "grad_norm": 0.013399476185441017,
+ "learning_rate": 0.0029197207362855735,
+ "loss": 5.890079021453857,
+ "step": 469
+ },
+ {
+ "epoch": 6.531004366812227,
+ "grad_norm": 0.025862423703074455,
+ "learning_rate": 0.0029192878443638427,
+ "loss": 5.919907569885254,
+ "step": 470
+ },
+ {
+ "epoch": 6.544978165938865,
+ "grad_norm": 0.011315986514091492,
+ "learning_rate": 0.0029188538210327595,
+ "loss": 5.901529788970947,
+ "step": 471
+ },
+ {
+ "epoch": 6.558951965065502,
+ "grad_norm": 0.01235947571694851,
+ "learning_rate": 0.00291841866664201,
+ "loss": 5.903595924377441,
+ "step": 472
+ },
+ {
+ "epoch": 6.5729257641921395,
+ "grad_norm": 0.012175127863883972,
+ "learning_rate": 0.0029179823815421893,
+ "loss": 5.868306636810303,
+ "step": 473
+ },
+ {
+ "epoch": 6.586899563318777,
+ "grad_norm": 0.01195445004850626,
+ "learning_rate": 0.0029175449660848052,
+ "loss": 5.899898529052734,
+ "step": 474
+ },
+ {
+ "epoch": 6.600873362445415,
+ "grad_norm": 0.012801481410861015,
+ "learning_rate": 0.0029171064206222745,
+ "loss": 5.890621662139893,
+ "step": 475
+ },
+ {
+ "epoch": 6.614847161572053,
+ "grad_norm": 0.011909718625247478,
+ "learning_rate": 0.002916666745507926,
+ "loss": 5.86421012878418,
+ "step": 476
+ },
+ {
+ "epoch": 6.62882096069869,
+ "grad_norm": 0.01162486057728529,
+ "learning_rate": 0.0029162259410959977,
+ "loss": 5.869139671325684,
+ "step": 477
+ },
+ {
+ "epoch": 6.642794759825327,
+ "grad_norm": 0.010923407040536404,
+ "learning_rate": 0.0029157840077416374,
+ "loss": 5.869789123535156,
+ "step": 478
+ },
+ {
+ "epoch": 6.656768558951965,
+ "grad_norm": 0.011943561024963856,
+ "learning_rate": 0.002915340945800904,
+ "loss": 5.873437881469727,
+ "step": 479
+ },
+ {
+ "epoch": 6.670742358078603,
+ "grad_norm": 0.010295520536601543,
+ "learning_rate": 0.002914896755630763,
+ "loss": 5.819953918457031,
+ "step": 480
+ },
+ {
+ "epoch": 6.68471615720524,
+ "grad_norm": 0.0093373479321599,
+ "learning_rate": 0.002914451437589092,
+ "loss": 5.829481601715088,
+ "step": 481
+ },
+ {
+ "epoch": 6.698689956331878,
+ "grad_norm": 0.01182442344725132,
+ "learning_rate": 0.002914004992034674,
+ "loss": 5.855249404907227,
+ "step": 482
+ },
+ {
+ "epoch": 6.712663755458515,
+ "grad_norm": 0.00939213763922453,
+ "learning_rate": 0.0029135574193272034,
+ "loss": 5.83876371383667,
+ "step": 483
+ },
+ {
+ "epoch": 6.726637554585153,
+ "grad_norm": 0.011648556217551231,
+ "learning_rate": 0.0029131087198272813,
+ "loss": 5.824906826019287,
+ "step": 484
+ },
+ {
+ "epoch": 6.74061135371179,
+ "grad_norm": 0.011227169074118137,
+ "learning_rate": 0.002912658893896416,
+ "loss": 5.847869873046875,
+ "step": 485
+ },
+ {
+ "epoch": 6.754585152838428,
+ "grad_norm": 0.011084851808845997,
+ "learning_rate": 0.002912207941897025,
+ "loss": 5.839183807373047,
+ "step": 486
+ },
+ {
+ "epoch": 6.7685589519650655,
+ "grad_norm": 0.009923825040459633,
+ "learning_rate": 0.002911755864192432,
+ "loss": 5.818821907043457,
+ "step": 487
+ },
+ {
+ "epoch": 6.782532751091703,
+ "grad_norm": 0.011300318874418736,
+ "learning_rate": 0.002911302661146867,
+ "loss": 5.819517135620117,
+ "step": 488
+ },
+ {
+ "epoch": 6.796506550218341,
+ "grad_norm": 0.009915645234286785,
+ "learning_rate": 0.0029108483331254684,
+ "loss": 5.776939392089844,
+ "step": 489
+ },
+ {
+ "epoch": 6.810480349344978,
+ "grad_norm": 0.008965205401182175,
+ "learning_rate": 0.00291039288049428,
+ "loss": 5.824148178100586,
+ "step": 490
+ },
+ {
+ "epoch": 6.824454148471616,
+ "grad_norm": 0.011025927029550076,
+ "learning_rate": 0.0029099363036202515,
+ "loss": 5.832704544067383,
+ "step": 491
+ },
+ {
+ "epoch": 6.8384279475982535,
+ "grad_norm": 0.010143411345779896,
+ "learning_rate": 0.0029094786028712387,
+ "loss": 5.802353382110596,
+ "step": 492
+ },
+ {
+ "epoch": 6.85240174672489,
+ "grad_norm": 0.010021256282925606,
+ "learning_rate": 0.002909019778616003,
+ "loss": 5.802136421203613,
+ "step": 493
+ },
+ {
+ "epoch": 6.866375545851528,
+ "grad_norm": 0.011356706731021404,
+ "learning_rate": 0.002908559831224211,
+ "loss": 5.807962417602539,
+ "step": 494
+ },
+ {
+ "epoch": 6.880349344978166,
+ "grad_norm": 0.010525298304855824,
+ "learning_rate": 0.002908098761066433,
+ "loss": 5.787661552429199,
+ "step": 495
+ },
+ {
+ "epoch": 6.894323144104804,
+ "grad_norm": 0.011147079057991505,
+ "learning_rate": 0.0029076365685141463,
+ "loss": 5.801015377044678,
+ "step": 496
+ },
+ {
+ "epoch": 6.908296943231441,
+ "grad_norm": 0.008777810260653496,
+ "learning_rate": 0.00290717325393973,
+ "loss": 5.779858589172363,
+ "step": 497
+ },
+ {
+ "epoch": 6.922270742358078,
+ "grad_norm": 0.011423609219491482,
+ "learning_rate": 0.002906708817716469,
+ "loss": 5.777166843414307,
+ "step": 498
+ },
+ {
+ "epoch": 6.936244541484716,
+ "grad_norm": 0.008639856241643429,
+ "learning_rate": 0.00290624326021855,
+ "loss": 5.747189521789551,
+ "step": 499
+ },
+ {
+ "epoch": 6.950218340611354,
+ "grad_norm": 0.009628898464143276,
+ "learning_rate": 0.002905776581821066,
+ "loss": 5.7666916847229,
+ "step": 500
+ },
+ {
+ "epoch": 6.964192139737992,
+ "grad_norm": 0.011988377198576927,
+ "learning_rate": 0.0029053087829000097,
+ "loss": 5.777951240539551,
+ "step": 501
+ },
+ {
+ "epoch": 6.978165938864628,
+ "grad_norm": 0.01693224348127842,
+ "learning_rate": 0.00290483986383228,
+ "loss": 5.818393707275391,
+ "step": 502
+ },
+ {
+ "epoch": 6.992139737991266,
+ "grad_norm": 0.020264796912670135,
+ "learning_rate": 0.002904369824995674,
+ "loss": 5.747611045837402,
+ "step": 503
+ },
+ {
+ "epoch": 7.0,
+ "grad_norm": 0.018283747136592865,
+ "learning_rate": 0.0029038986667688964,
+ "loss": 5.789663314819336,
+ "step": 504
+ },
+ {
+ "epoch": 7.0,
+ "eval_loss": 5.825676918029785,
+ "eval_runtime": 57.6908,
+ "eval_samples_per_second": 42.329,
+ "eval_steps_per_second": 1.335,
+ "step": 504
+ },
+ {
+ "epoch": 7.013973799126638,
+ "grad_norm": 0.011447477154433727,
+ "learning_rate": 0.002903426389531549,
+ "loss": 5.724975109100342,
+ "step": 505
+ },
+ {
+ "epoch": 7.0279475982532755,
+ "grad_norm": 0.012539781630039215,
+ "learning_rate": 0.002902952993664138,
+ "loss": 5.747196674346924,
+ "step": 506
+ },
+ {
+ "epoch": 7.041921397379912,
+ "grad_norm": 0.012179837562143803,
+ "learning_rate": 0.002902478479548069,
+ "loss": 5.748995780944824,
+ "step": 507
+ },
+ {
+ "epoch": 7.05589519650655,
+ "grad_norm": 0.019896456971764565,
+ "learning_rate": 0.00290200284756565,
+ "loss": 5.751996994018555,
+ "step": 508
+ },
+ {
+ "epoch": 7.069868995633188,
+ "grad_norm": 0.022261742502450943,
+ "learning_rate": 0.002901526098100089,
+ "loss": 5.7516984939575195,
+ "step": 509
+ },
+ {
+ "epoch": 7.083842794759826,
+ "grad_norm": 0.02683536894619465,
+ "learning_rate": 0.0029010482315354946,
+ "loss": 5.716037750244141,
+ "step": 510
+ },
+ {
+ "epoch": 7.097816593886463,
+ "grad_norm": 0.018057072535157204,
+ "learning_rate": 0.0029005692482568747,
+ "loss": 5.778608322143555,
+ "step": 511
+ },
+ {
+ "epoch": 7.1117903930131,
+ "grad_norm": 0.011646891944110394,
+ "learning_rate": 0.002900089148650138,
+ "loss": 5.7308268547058105,
+ "step": 512
+ },
+ {
+ "epoch": 7.125764192139738,
+ "grad_norm": 0.02230495773255825,
+ "learning_rate": 0.002899607933102092,
+ "loss": 5.768552780151367,
+ "step": 513
+ },
+ {
+ "epoch": 7.139737991266376,
+ "grad_norm": 0.01212464738637209,
+ "learning_rate": 0.002899125602000443,
+ "loss": 5.733015060424805,
+ "step": 514
+ },
+ {
+ "epoch": 7.153711790393013,
+ "grad_norm": 0.011764907278120518,
+ "learning_rate": 0.0028986421557337973,
+ "loss": 5.748104095458984,
+ "step": 515
+ },
+ {
+ "epoch": 7.1676855895196505,
+ "grad_norm": 0.01593147963285446,
+ "learning_rate": 0.002898157594691658,
+ "loss": 5.701960563659668,
+ "step": 516
+ },
+ {
+ "epoch": 7.181659388646288,
+ "grad_norm": 0.016333773732185364,
+ "learning_rate": 0.0028976719192644277,
+ "loss": 5.716808795928955,
+ "step": 517
+ },
+ {
+ "epoch": 7.195633187772926,
+ "grad_norm": 0.017547868192195892,
+ "learning_rate": 0.002897185129843406,
+ "loss": 5.724377632141113,
+ "step": 518
+ },
+ {
+ "epoch": 7.209606986899563,
+ "grad_norm": 0.01662801019847393,
+ "learning_rate": 0.002896697226820791,
+ "loss": 5.737432956695557,
+ "step": 519
+ },
+ {
+ "epoch": 7.223580786026201,
+ "grad_norm": 0.015781691297888756,
+ "learning_rate": 0.002896208210589676,
+ "loss": 5.7376203536987305,
+ "step": 520
+ },
+ {
+ "epoch": 7.2375545851528384,
+ "grad_norm": 0.014363214373588562,
+ "learning_rate": 0.0028957180815440534,
+ "loss": 5.7342610359191895,
+ "step": 521
+ },
+ {
+ "epoch": 7.251528384279476,
+ "grad_norm": 0.015048054046928883,
+ "learning_rate": 0.0028952268400788115,
+ "loss": 5.715073585510254,
+ "step": 522
+ },
+ {
+ "epoch": 7.265502183406113,
+ "grad_norm": 0.013753226026892662,
+ "learning_rate": 0.002894734486589734,
+ "loss": 5.73264217376709,
+ "step": 523
+ },
+ {
+ "epoch": 7.279475982532751,
+ "grad_norm": 0.015562563203275204,
+ "learning_rate": 0.002894241021473502,
+ "loss": 5.7072038650512695,
+ "step": 524
+ },
+ {
+ "epoch": 7.293449781659389,
+ "grad_norm": 0.01520819216966629,
+ "learning_rate": 0.0028937464451276916,
+ "loss": 5.683300018310547,
+ "step": 525
+ },
+ {
+ "epoch": 7.307423580786026,
+ "grad_norm": 0.013024724088609219,
+ "learning_rate": 0.002893250757950773,
+ "loss": 5.726273536682129,
+ "step": 526
+ },
+ {
+ "epoch": 7.321397379912664,
+ "grad_norm": 0.012760856188833714,
+ "learning_rate": 0.0028927539603421128,
+ "loss": 5.7249064445495605,
+ "step": 527
+ },
+ {
+ "epoch": 7.335371179039301,
+ "grad_norm": 0.013542579486966133,
+ "learning_rate": 0.0028922560527019723,
+ "loss": 5.70849084854126,
+ "step": 528
+ },
+ {
+ "epoch": 7.349344978165939,
+ "grad_norm": 0.014801754616200924,
+ "learning_rate": 0.002891757035431506,
+ "loss": 5.725185394287109,
+ "step": 529
+ },
+ {
+ "epoch": 7.3633187772925766,
+ "grad_norm": 0.013000023551285267,
+ "learning_rate": 0.0028912569089327637,
+ "loss": 5.705799102783203,
+ "step": 530
+ },
+ {
+ "epoch": 7.377292576419214,
+ "grad_norm": 0.012919292785227299,
+ "learning_rate": 0.002890755673608688,
+ "loss": 5.674324989318848,
+ "step": 531
+ },
+ {
+ "epoch": 7.391266375545851,
+ "grad_norm": 0.015904201194643974,
+ "learning_rate": 0.002890253329863115,
+ "loss": 5.680788993835449,
+ "step": 532
+ },
+ {
+ "epoch": 7.405240174672489,
+ "grad_norm": 0.015467721968889236,
+ "learning_rate": 0.0028897498781007744,
+ "loss": 5.707531929016113,
+ "step": 533
+ },
+ {
+ "epoch": 7.419213973799127,
+ "grad_norm": 0.015157260000705719,
+ "learning_rate": 0.002889245318727288,
+ "loss": 5.689894676208496,
+ "step": 534
+ },
+ {
+ "epoch": 7.4331877729257645,
+ "grad_norm": 0.014204306527972221,
+ "learning_rate": 0.00288873965214917,
+ "loss": 5.713001251220703,
+ "step": 535
+ },
+ {
+ "epoch": 7.447161572052401,
+ "grad_norm": 0.022045845165848732,
+ "learning_rate": 0.0028882328787738273,
+ "loss": 5.657631874084473,
+ "step": 536
+ },
+ {
+ "epoch": 7.461135371179039,
+ "grad_norm": 0.013380175456404686,
+ "learning_rate": 0.0028877249990095573,
+ "loss": 5.702192306518555,
+ "step": 537
+ },
+ {
+ "epoch": 7.475109170305677,
+ "grad_norm": 0.016698408871889114,
+ "learning_rate": 0.0028872160132655503,
+ "loss": 5.695206642150879,
+ "step": 538
+ },
+ {
+ "epoch": 7.489082969432315,
+ "grad_norm": 0.020001472905278206,
+ "learning_rate": 0.0028867059219518866,
+ "loss": 5.731742858886719,
+ "step": 539
+ },
+ {
+ "epoch": 7.503056768558952,
+ "grad_norm": 0.027350103482604027,
+ "learning_rate": 0.0028861947254795375,
+ "loss": 5.712314605712891,
+ "step": 540
+ },
+ {
+ "epoch": 7.517030567685589,
+ "grad_norm": 0.021862946450710297,
+ "learning_rate": 0.002885682424260365,
+ "loss": 5.697043418884277,
+ "step": 541
+ },
+ {
+ "epoch": 7.531004366812227,
+ "grad_norm": 0.019216977059841156,
+ "learning_rate": 0.0028851690187071204,
+ "loss": 5.698813438415527,
+ "step": 542
+ },
+ {
+ "epoch": 7.544978165938865,
+ "grad_norm": 0.02131095714867115,
+ "learning_rate": 0.002884654509233446,
+ "loss": 5.6918864250183105,
+ "step": 543
+ },
+ {
+ "epoch": 7.558951965065502,
+ "grad_norm": 0.016583887860178947,
+ "learning_rate": 0.0028841388962538724,
+ "loss": 5.711456298828125,
+ "step": 544
+ },
+ {
+ "epoch": 7.5729257641921395,
+ "grad_norm": 0.016998684033751488,
+ "learning_rate": 0.0028836221801838194,
+ "loss": 5.675747871398926,
+ "step": 545
+ },
+ {
+ "epoch": 7.586899563318777,
+ "grad_norm": 0.01705755852162838,
+ "learning_rate": 0.0028831043614395966,
+ "loss": 5.697177410125732,
+ "step": 546
+ },
+ {
+ "epoch": 7.600873362445415,
+ "grad_norm": 0.014604352414608002,
+ "learning_rate": 0.0028825854404384,
+ "loss": 5.681950092315674,
+ "step": 547
+ },
+ {
+ "epoch": 7.614847161572053,
+ "grad_norm": 0.013561096973717213,
+ "learning_rate": 0.0028820654175983162,
+ "loss": 5.67357063293457,
+ "step": 548
+ },
+ {
+ "epoch": 7.62882096069869,
+ "grad_norm": 0.021149173378944397,
+ "learning_rate": 0.0028815442933383174,
+ "loss": 5.709477424621582,
+ "step": 549
+ },
+ {
+ "epoch": 7.642794759825327,
+ "grad_norm": 0.01431041955947876,
+ "learning_rate": 0.0028810220680782644,
+ "loss": 5.7012200355529785,
+ "step": 550
+ },
+ {
+ "epoch": 7.656768558951965,
+ "grad_norm": 0.01429709978401661,
+ "learning_rate": 0.0028804987422389045,
+ "loss": 5.689794540405273,
+ "step": 551
+ },
+ {
+ "epoch": 7.670742358078603,
+ "grad_norm": 0.032025713473558426,
+ "learning_rate": 0.0028799743162418712,
+ "loss": 5.6713762283325195,
+ "step": 552
+ },
+ {
+ "epoch": 7.68471615720524,
+ "grad_norm": 0.013168714940547943,
+ "learning_rate": 0.002879448790509686,
+ "loss": 5.685478210449219,
+ "step": 553
+ },
+ {
+ "epoch": 7.698689956331878,
+ "grad_norm": 0.014088666997849941,
+ "learning_rate": 0.0028789221654657556,
+ "loss": 5.645514965057373,
+ "step": 554
+ },
+ {
+ "epoch": 7.712663755458515,
+ "grad_norm": 0.013320053927600384,
+ "learning_rate": 0.002878394441534372,
+ "loss": 5.678678512573242,
+ "step": 555
+ },
+ {
+ "epoch": 7.726637554585153,
+ "grad_norm": 0.022180087864398956,
+ "learning_rate": 0.002877865619140712,
+ "loss": 5.660525321960449,
+ "step": 556
+ },
+ {
+ "epoch": 7.74061135371179,
+ "grad_norm": 0.01271889265626669,
+ "learning_rate": 0.002877335698710839,
+ "loss": 5.652164459228516,
+ "step": 557
+ },
+ {
+ "epoch": 7.754585152838428,
+ "grad_norm": 0.011617431417107582,
+ "learning_rate": 0.002876804680671701,
+ "loss": 5.6389946937561035,
+ "step": 558
+ },
+ {
+ "epoch": 7.7685589519650655,
+ "grad_norm": 0.012040726840496063,
+ "learning_rate": 0.0028762725654511282,
+ "loss": 5.628744602203369,
+ "step": 559
+ },
+ {
+ "epoch": 7.782532751091703,
+ "grad_norm": 0.013148069381713867,
+ "learning_rate": 0.0028757393534778377,
+ "loss": 5.69610595703125,
+ "step": 560
+ },
+ {
+ "epoch": 7.796506550218341,
+ "grad_norm": 0.01826276071369648,
+ "learning_rate": 0.0028752050451814276,
+ "loss": 5.634238243103027,
+ "step": 561
+ },
+ {
+ "epoch": 7.810480349344978,
+ "grad_norm": 0.020393460988998413,
+ "learning_rate": 0.0028746696409923805,
+ "loss": 5.656968593597412,
+ "step": 562
+ },
+ {
+ "epoch": 7.824454148471616,
+ "grad_norm": 0.017072148621082306,
+ "learning_rate": 0.0028741331413420627,
+ "loss": 5.668219089508057,
+ "step": 563
+ },
+ {
+ "epoch": 7.8384279475982535,
+ "grad_norm": 0.01694098301231861,
+ "learning_rate": 0.0028735955466627217,
+ "loss": 5.670073509216309,
+ "step": 564
+ },
+ {
+ "epoch": 7.85240174672489,
+ "grad_norm": 0.022760670632123947,
+ "learning_rate": 0.002873056857387488,
+ "loss": 5.649288177490234,
+ "step": 565
+ },
+ {
+ "epoch": 7.866375545851528,
+ "grad_norm": 0.03613442927598953,
+ "learning_rate": 0.0028725170739503734,
+ "loss": 5.723211288452148,
+ "step": 566
+ },
+ {
+ "epoch": 7.880349344978166,
+ "grad_norm": 0.04232329502701759,
+ "learning_rate": 0.002871976196786272,
+ "loss": 5.757920742034912,
+ "step": 567
+ },
+ {
+ "epoch": 7.894323144104804,
+ "grad_norm": 0.044210851192474365,
+ "learning_rate": 0.002871434226330959,
+ "loss": 5.702426910400391,
+ "step": 568
+ },
+ {
+ "epoch": 7.908296943231441,
+ "grad_norm": 0.044886402785778046,
+ "learning_rate": 0.0028708911630210905,
+ "loss": 5.742881774902344,
+ "step": 569
+ },
+ {
+ "epoch": 7.922270742358078,
+ "grad_norm": 0.022805195301771164,
+ "learning_rate": 0.002870347007294202,
+ "loss": 5.716353416442871,
+ "step": 570
+ },
+ {
+ "epoch": 7.936244541484716,
+ "grad_norm": 0.02219395525753498,
+ "learning_rate": 0.00286980175958871,
+ "loss": 5.68503475189209,
+ "step": 571
+ },
+ {
+ "epoch": 7.950218340611354,
+ "grad_norm": 0.019240057095885277,
+ "learning_rate": 0.002869255420343911,
+ "loss": 5.700656890869141,
+ "step": 572
+ },
+ {
+ "epoch": 7.964192139737992,
+ "grad_norm": 0.01957295648753643,
+ "learning_rate": 0.0028687079899999807,
+ "loss": 5.70595645904541,
+ "step": 573
+ },
+ {
+ "epoch": 7.978165938864628,
+ "grad_norm": 0.016326555982232094,
+ "learning_rate": 0.0028681594689979736,
+ "loss": 5.697229385375977,
+ "step": 574
+ },
+ {
+ "epoch": 7.992139737991266,
+ "grad_norm": 0.015815885737538338,
+ "learning_rate": 0.0028676098577798234,
+ "loss": 5.7020792961120605,
+ "step": 575
+ },
+ {
+ "epoch": 8.0,
+ "grad_norm": 0.015865175053477287,
+ "learning_rate": 0.002867059156788342,
+ "loss": 5.648272514343262,
+ "step": 576
+ },
+ {
+ "epoch": 8.0,
+ "eval_loss": 5.7417449951171875,
+ "eval_runtime": 51.2263,
+ "eval_samples_per_second": 47.671,
+ "eval_steps_per_second": 1.503,
+ "step": 576
+ },
+ {
+ "epoch": 8.013973799126637,
+ "grad_norm": 0.016063183546066284,
+ "learning_rate": 0.0028665073664672185,
+ "loss": 5.65411376953125,
+ "step": 577
+ },
+ {
+ "epoch": 8.027947598253276,
+ "grad_norm": 0.014323881827294827,
+ "learning_rate": 0.0028659544872610214,
+ "loss": 5.629649639129639,
+ "step": 578
+ },
+ {
+ "epoch": 8.041921397379912,
+ "grad_norm": 0.014005238190293312,
+ "learning_rate": 0.002865400519615195,
+ "loss": 5.655135631561279,
+ "step": 579
+ },
+ {
+ "epoch": 8.055895196506551,
+ "grad_norm": 0.01385372132062912,
+ "learning_rate": 0.0028648454639760615,
+ "loss": 5.632213115692139,
+ "step": 580
+ },
+ {
+ "epoch": 8.069868995633188,
+ "grad_norm": 0.014073756523430347,
+ "learning_rate": 0.0028642893207908187,
+ "loss": 5.626431465148926,
+ "step": 581
+ },
+ {
+ "epoch": 8.083842794759825,
+ "grad_norm": 0.019551943987607956,
+ "learning_rate": 0.0028637320905075406,
+ "loss": 5.595447063446045,
+ "step": 582
+ },
+ {
+ "epoch": 8.097816593886463,
+ "grad_norm": 0.017571093514561653,
+ "learning_rate": 0.002863173773575179,
+ "loss": 5.661581039428711,
+ "step": 583
+ },
+ {
+ "epoch": 8.1117903930131,
+ "grad_norm": 0.01691303588449955,
+ "learning_rate": 0.0028626143704435585,
+ "loss": 5.641129493713379,
+ "step": 584
+ },
+ {
+ "epoch": 8.125764192139737,
+ "grad_norm": 0.01584242656826973,
+ "learning_rate": 0.0028620538815633807,
+ "loss": 5.640626907348633,
+ "step": 585
+ },
+ {
+ "epoch": 8.139737991266376,
+ "grad_norm": 0.014592235907912254,
+ "learning_rate": 0.0028614923073862213,
+ "loss": 5.611204147338867,
+ "step": 586
+ },
+ {
+ "epoch": 8.153711790393013,
+ "grad_norm": 0.016636405140161514,
+ "learning_rate": 0.0028609296483645304,
+ "loss": 5.614716053009033,
+ "step": 587
+ },
+ {
+ "epoch": 8.167685589519651,
+ "grad_norm": 0.01370833721011877,
+ "learning_rate": 0.002860365904951632,
+ "loss": 5.633011817932129,
+ "step": 588
+ },
+ {
+ "epoch": 8.181659388646288,
+ "grad_norm": 0.013218311592936516,
+ "learning_rate": 0.002859801077601725,
+ "loss": 5.588958740234375,
+ "step": 589
+ },
+ {
+ "epoch": 8.195633187772925,
+ "grad_norm": 0.016971616074442863,
+ "learning_rate": 0.0028592351667698786,
+ "loss": 5.635064125061035,
+ "step": 590
+ },
+ {
+ "epoch": 8.209606986899564,
+ "grad_norm": 0.015524905174970627,
+ "learning_rate": 0.0028586681729120393,
+ "loss": 5.611454963684082,
+ "step": 591
+ },
+ {
+ "epoch": 8.2235807860262,
+ "grad_norm": 0.014769157394766808,
+ "learning_rate": 0.0028581000964850215,
+ "loss": 5.62175178527832,
+ "step": 592
+ },
+ {
+ "epoch": 8.237554585152838,
+ "grad_norm": 0.014313343912363052,
+ "learning_rate": 0.002857530937946516,
+ "loss": 5.603628158569336,
+ "step": 593
+ },
+ {
+ "epoch": 8.251528384279476,
+ "grad_norm": 0.016748322173953056,
+ "learning_rate": 0.002856960697755083,
+ "loss": 5.617765426635742,
+ "step": 594
+ },
+ {
+ "epoch": 8.265502183406113,
+ "grad_norm": 0.013503300957381725,
+ "learning_rate": 0.0028563893763701538,
+ "loss": 5.636135578155518,
+ "step": 595
+ },
+ {
+ "epoch": 8.279475982532752,
+ "grad_norm": 0.012275655753910542,
+ "learning_rate": 0.002855816974252033,
+ "loss": 5.552533149719238,
+ "step": 596
+ },
+ {
+ "epoch": 8.293449781659389,
+ "grad_norm": 0.01288773212581873,
+ "learning_rate": 0.002855243491861894,
+ "loss": 5.620527267456055,
+ "step": 597
+ },
+ {
+ "epoch": 8.307423580786025,
+ "grad_norm": 0.013687008060514927,
+ "learning_rate": 0.002854668929661781,
+ "loss": 5.618595123291016,
+ "step": 598
+ },
+ {
+ "epoch": 8.321397379912664,
+ "grad_norm": 0.015678439289331436,
+ "learning_rate": 0.0028540932881146092,
+ "loss": 5.604031562805176,
+ "step": 599
+ },
+ {
+ "epoch": 8.335371179039301,
+ "grad_norm": 0.013567988760769367,
+ "learning_rate": 0.002853516567684162,
+ "loss": 5.549149513244629,
+ "step": 600
+ },
+ {
+ "epoch": 8.34934497816594,
+ "grad_norm": 0.0157297495752573,
+ "learning_rate": 0.002852938768835092,
+ "loss": 5.618505477905273,
+ "step": 601
+ },
+ {
+ "epoch": 8.363318777292577,
+ "grad_norm": 0.011867019347846508,
+ "learning_rate": 0.0028523598920329233,
+ "loss": 5.579049110412598,
+ "step": 602
+ },
+ {
+ "epoch": 8.377292576419213,
+ "grad_norm": 0.01398563478142023,
+ "learning_rate": 0.0028517799377440443,
+ "loss": 5.597678184509277,
+ "step": 603
+ },
+ {
+ "epoch": 8.391266375545852,
+ "grad_norm": 0.015525273978710175,
+ "learning_rate": 0.0028511989064357148,
+ "loss": 5.601663589477539,
+ "step": 604
+ },
+ {
+ "epoch": 8.405240174672489,
+ "grad_norm": 0.01588466763496399,
+ "learning_rate": 0.002850616798576061,
+ "loss": 5.5650434494018555,
+ "step": 605
+ },
+ {
+ "epoch": 8.419213973799126,
+ "grad_norm": 0.014276969246566296,
+ "learning_rate": 0.0028500336146340772,
+ "loss": 5.607125282287598,
+ "step": 606
+ },
+ {
+ "epoch": 8.433187772925764,
+ "grad_norm": 0.013253689743578434,
+ "learning_rate": 0.002849449355079623,
+ "loss": 5.6299824714660645,
+ "step": 607
+ },
+ {
+ "epoch": 8.447161572052401,
+ "grad_norm": 0.01176661066710949,
+ "learning_rate": 0.0028488640203834263,
+ "loss": 5.560839653015137,
+ "step": 608
+ },
+ {
+ "epoch": 8.46113537117904,
+ "grad_norm": 0.011712565086781979,
+ "learning_rate": 0.0028482776110170813,
+ "loss": 5.618312835693359,
+ "step": 609
+ },
+ {
+ "epoch": 8.475109170305677,
+ "grad_norm": 0.013515647500753403,
+ "learning_rate": 0.002847690127453047,
+ "loss": 5.566764831542969,
+ "step": 610
+ },
+ {
+ "epoch": 8.489082969432314,
+ "grad_norm": 0.020177524536848068,
+ "learning_rate": 0.0028471015701646485,
+ "loss": 5.554571628570557,
+ "step": 611
+ },
+ {
+ "epoch": 8.503056768558952,
+ "grad_norm": 0.029197178781032562,
+ "learning_rate": 0.0028465119396260757,
+ "loss": 5.6097259521484375,
+ "step": 612
+ },
+ {
+ "epoch": 8.51703056768559,
+ "grad_norm": 0.027947140857577324,
+ "learning_rate": 0.0028459212363123824,
+ "loss": 5.65598201751709,
+ "step": 613
+ },
+ {
+ "epoch": 8.531004366812226,
+ "grad_norm": 0.021089809015393257,
+ "learning_rate": 0.0028453294606994887,
+ "loss": 5.67195987701416,
+ "step": 614
+ },
+ {
+ "epoch": 8.544978165938865,
+ "grad_norm": 0.019936271011829376,
+ "learning_rate": 0.0028447366132641773,
+ "loss": 5.609482765197754,
+ "step": 615
+ },
+ {
+ "epoch": 8.558951965065502,
+ "grad_norm": 0.0176444910466671,
+ "learning_rate": 0.002844142694484094,
+ "loss": 5.60432767868042,
+ "step": 616
+ },
+ {
+ "epoch": 8.57292576419214,
+ "grad_norm": 0.023154428228735924,
+ "learning_rate": 0.0028435477048377495,
+ "loss": 5.616270065307617,
+ "step": 617
+ },
+ {
+ "epoch": 8.586899563318777,
+ "grad_norm": 0.018391480669379234,
+ "learning_rate": 0.002842951644804515,
+ "loss": 5.600588321685791,
+ "step": 618
+ },
+ {
+ "epoch": 8.600873362445414,
+ "grad_norm": 0.013716557994484901,
+ "learning_rate": 0.002842354514864627,
+ "loss": 5.595095634460449,
+ "step": 619
+ },
+ {
+ "epoch": 8.614847161572053,
+ "grad_norm": 0.014372860081493855,
+ "learning_rate": 0.002841756315499181,
+ "loss": 5.584517478942871,
+ "step": 620
+ },
+ {
+ "epoch": 8.62882096069869,
+ "grad_norm": 0.014930845238268375,
+ "learning_rate": 0.002841157047190135,
+ "loss": 5.587005138397217,
+ "step": 621
+ },
+ {
+ "epoch": 8.642794759825328,
+ "grad_norm": 0.012631502002477646,
+ "learning_rate": 0.0028405567104203095,
+ "loss": 5.594094276428223,
+ "step": 622
+ },
+ {
+ "epoch": 8.656768558951965,
+ "grad_norm": 0.014677002094686031,
+ "learning_rate": 0.0028399553056733854,
+ "loss": 5.569582939147949,
+ "step": 623
+ },
+ {
+ "epoch": 8.670742358078602,
+ "grad_norm": 0.015558961778879166,
+ "learning_rate": 0.0028393528334339025,
+ "loss": 5.596535682678223,
+ "step": 624
+ },
+ {
+ "epoch": 8.68471615720524,
+ "grad_norm": 0.014322095550596714,
+ "learning_rate": 0.0028387492941872624,
+ "loss": 5.570324420928955,
+ "step": 625
+ },
+ {
+ "epoch": 8.698689956331878,
+ "grad_norm": 0.012913414277136326,
+ "learning_rate": 0.002838144688419726,
+ "loss": 5.573002338409424,
+ "step": 626
+ },
+ {
+ "epoch": 8.712663755458514,
+ "grad_norm": 0.01251310296356678,
+ "learning_rate": 0.0028375390166184137,
+ "loss": 5.550327777862549,
+ "step": 627
+ },
+ {
+ "epoch": 8.726637554585153,
+ "grad_norm": 0.013847017660737038,
+ "learning_rate": 0.0028369322792713026,
+ "loss": 5.54567813873291,
+ "step": 628
+ },
+ {
+ "epoch": 8.74061135371179,
+ "grad_norm": 0.013789358548820019,
+ "learning_rate": 0.0028363244768672313,
+ "loss": 5.571495056152344,
+ "step": 629
+ },
+ {
+ "epoch": 8.754585152838429,
+ "grad_norm": 0.01236631814390421,
+ "learning_rate": 0.0028357156098958954,
+ "loss": 5.569730758666992,
+ "step": 630
+ },
+ {
+ "epoch": 8.768558951965066,
+ "grad_norm": 0.013210169039666653,
+ "learning_rate": 0.002835105678847847,
+ "loss": 5.53984260559082,
+ "step": 631
+ },
+ {
+ "epoch": 8.782532751091702,
+ "grad_norm": 0.01604137383401394,
+ "learning_rate": 0.0028344946842144975,
+ "loss": 5.530598163604736,
+ "step": 632
+ },
+ {
+ "epoch": 8.796506550218341,
+ "grad_norm": 0.01708815060555935,
+ "learning_rate": 0.002833882626488114,
+ "loss": 5.529256820678711,
+ "step": 633
+ },
+ {
+ "epoch": 8.810480349344978,
+ "grad_norm": 0.017222166061401367,
+ "learning_rate": 0.0028332695061618197,
+ "loss": 5.564949035644531,
+ "step": 634
+ },
+ {
+ "epoch": 8.824454148471617,
+ "grad_norm": 0.01803218014538288,
+ "learning_rate": 0.002832655323729595,
+ "loss": 5.560049057006836,
+ "step": 635
+ },
+ {
+ "epoch": 8.838427947598253,
+ "grad_norm": 0.01658225618302822,
+ "learning_rate": 0.002832040079686276,
+ "loss": 5.553183555603027,
+ "step": 636
+ },
+ {
+ "epoch": 8.85240174672489,
+ "grad_norm": 0.016834143549203873,
+ "learning_rate": 0.002831423774527554,
+ "loss": 5.573678970336914,
+ "step": 637
+ },
+ {
+ "epoch": 8.866375545851529,
+ "grad_norm": 0.01856408454477787,
+ "learning_rate": 0.002830806408749974,
+ "loss": 5.539497375488281,
+ "step": 638
+ },
+ {
+ "epoch": 8.880349344978166,
+ "grad_norm": 0.026909375563263893,
+ "learning_rate": 0.002830187982850936,
+ "loss": 5.551433563232422,
+ "step": 639
+ },
+ {
+ "epoch": 8.894323144104803,
+ "grad_norm": 0.016046348959207535,
+ "learning_rate": 0.002829568497328696,
+ "loss": 5.566728591918945,
+ "step": 640
+ },
+ {
+ "epoch": 8.908296943231441,
+ "grad_norm": 0.018582617864012718,
+ "learning_rate": 0.0028289479526823614,
+ "loss": 5.543226718902588,
+ "step": 641
+ },
+ {
+ "epoch": 8.922270742358078,
+ "grad_norm": 0.0211537666618824,
+ "learning_rate": 0.0028283263494118943,
+ "loss": 5.560711860656738,
+ "step": 642
+ },
+ {
+ "epoch": 8.936244541484717,
+ "grad_norm": 0.013894101604819298,
+ "learning_rate": 0.0028277036880181095,
+ "loss": 5.528890132904053,
+ "step": 643
+ },
+ {
+ "epoch": 8.950218340611354,
+ "grad_norm": 0.017581885680556297,
+ "learning_rate": 0.0028270799690026725,
+ "loss": 5.57473087310791,
+ "step": 644
+ },
+ {
+ "epoch": 8.96419213973799,
+ "grad_norm": 0.01622696965932846,
+ "learning_rate": 0.002826455192868104,
+ "loss": 5.578961372375488,
+ "step": 645
+ },
+ {
+ "epoch": 8.97816593886463,
+ "grad_norm": 0.01966337114572525,
+ "learning_rate": 0.0028258293601177743,
+ "loss": 5.513942718505859,
+ "step": 646
+ },
+ {
+ "epoch": 8.992139737991266,
+ "grad_norm": 0.021581366658210754,
+ "learning_rate": 0.002825202471255906,
+ "loss": 5.573669910430908,
+ "step": 647
+ },
+ {
+ "epoch": 9.0,
+ "grad_norm": 0.020468758419156075,
+ "learning_rate": 0.002824574526787571,
+ "loss": 5.516103744506836,
+ "step": 648
+ },
+ {
+ "epoch": 9.0,
+ "eval_loss": 5.626922130584717,
+ "eval_runtime": 54.2404,
+ "eval_samples_per_second": 45.022,
+ "eval_steps_per_second": 1.42,
+ "step": 648
+ },
+ {
+ "epoch": 9.013973799126637,
+ "grad_norm": 0.014597049914300442,
+ "learning_rate": 0.0028239455272186936,
+ "loss": 5.506453514099121,
+ "step": 649
+ },
+ {
+ "epoch": 9.027947598253276,
+ "grad_norm": 0.020277054980397224,
+ "learning_rate": 0.002823315473056048,
+ "loss": 5.535350799560547,
+ "step": 650
+ },
+ {
+ "epoch": 9.041921397379912,
+ "grad_norm": 0.019103925675153732,
+ "learning_rate": 0.0028226843648072562,
+ "loss": 5.522078037261963,
+ "step": 651
+ },
+ {
+ "epoch": 9.055895196506551,
+ "grad_norm": 0.07158710062503815,
+ "learning_rate": 0.0028220522029807913,
+ "loss": 5.531142234802246,
+ "step": 652
+ },
+ {
+ "epoch": 9.069868995633188,
+ "grad_norm": 0.01958007924258709,
+ "learning_rate": 0.0028214189880859746,
+ "loss": 5.484055042266846,
+ "step": 653
+ },
+ {
+ "epoch": 9.083842794759825,
+ "grad_norm": 0.0185184795409441,
+ "learning_rate": 0.002820784720632976,
+ "loss": 5.511264801025391,
+ "step": 654
+ },
+ {
+ "epoch": 9.097816593886463,
+ "grad_norm": 0.01976991444826126,
+ "learning_rate": 0.0028201494011328133,
+ "loss": 5.51821231842041,
+ "step": 655
+ },
+ {
+ "epoch": 9.1117903930131,
+ "grad_norm": 0.02188604325056076,
+ "learning_rate": 0.002819513030097352,
+ "loss": 5.483095169067383,
+ "step": 656
+ },
+ {
+ "epoch": 9.125764192139737,
+ "grad_norm": 0.019405223429203033,
+ "learning_rate": 0.002818875608039305,
+ "loss": 5.539701461791992,
+ "step": 657
+ },
+ {
+ "epoch": 9.139737991266376,
+ "grad_norm": 0.016907615587115288,
+ "learning_rate": 0.0028182371354722306,
+ "loss": 5.518251419067383,
+ "step": 658
+ },
+ {
+ "epoch": 9.153711790393013,
+ "grad_norm": 0.02160538360476494,
+ "learning_rate": 0.0028175976129105364,
+ "loss": 5.509786605834961,
+ "step": 659
+ },
+ {
+ "epoch": 9.167685589519651,
+ "grad_norm": 0.023125721141695976,
+ "learning_rate": 0.002816957040869473,
+ "loss": 5.535561561584473,
+ "step": 660
+ },
+ {
+ "epoch": 9.181659388646288,
+ "grad_norm": 0.022233380004763603,
+ "learning_rate": 0.0028163154198651372,
+ "loss": 5.553831100463867,
+ "step": 661
+ },
+ {
+ "epoch": 9.195633187772925,
+ "grad_norm": 0.017227552831172943,
+ "learning_rate": 0.002815672750414473,
+ "loss": 5.547080993652344,
+ "step": 662
+ },
+ {
+ "epoch": 9.209606986899564,
+ "grad_norm": 0.017122473567724228,
+ "learning_rate": 0.0028150290330352665,
+ "loss": 5.493778228759766,
+ "step": 663
+ },
+ {
+ "epoch": 9.2235807860262,
+ "grad_norm": 0.020641913637518883,
+ "learning_rate": 0.0028143842682461498,
+ "loss": 5.536645889282227,
+ "step": 664
+ },
+ {
+ "epoch": 9.237554585152838,
+ "grad_norm": 0.017383966594934464,
+ "learning_rate": 0.0028137384565665967,
+ "loss": 5.506233215332031,
+ "step": 665
+ },
+ {
+ "epoch": 9.251528384279476,
+ "grad_norm": 0.017969606444239616,
+ "learning_rate": 0.002813091598516928,
+ "loss": 5.524741172790527,
+ "step": 666
+ },
+ {
+ "epoch": 9.265502183406113,
+ "grad_norm": 0.01936613954603672,
+ "learning_rate": 0.0028124436946183047,
+ "loss": 5.460989952087402,
+ "step": 667
+ },
+ {
+ "epoch": 9.279475982532752,
+ "grad_norm": 0.017492495477199554,
+ "learning_rate": 0.0028117947453927313,
+ "loss": 5.487619876861572,
+ "step": 668
+ },
+ {
+ "epoch": 9.293449781659389,
+ "grad_norm": 0.018786396831274033,
+ "learning_rate": 0.0028111447513630545,
+ "loss": 5.516829013824463,
+ "step": 669
+ },
+ {
+ "epoch": 9.307423580786025,
+ "grad_norm": 0.020149117335677147,
+ "learning_rate": 0.0028104937130529624,
+ "loss": 5.460541725158691,
+ "step": 670
+ },
+ {
+ "epoch": 9.321397379912664,
+ "grad_norm": 0.024668103083968163,
+ "learning_rate": 0.0028098416309869857,
+ "loss": 5.497687339782715,
+ "step": 671
+ },
+ {
+ "epoch": 9.335371179039301,
+ "grad_norm": 0.015888608992099762,
+ "learning_rate": 0.002809188505690494,
+ "loss": 5.494459629058838,
+ "step": 672
+ },
+ {
+ "epoch": 9.34934497816594,
+ "grad_norm": 0.019327623769640923,
+ "learning_rate": 0.0028085343376896997,
+ "loss": 5.486351490020752,
+ "step": 673
+ },
+ {
+ "epoch": 9.363318777292577,
+ "grad_norm": 0.026855193078517914,
+ "learning_rate": 0.0028078791275116532,
+ "loss": 5.509825706481934,
+ "step": 674
+ },
+ {
+ "epoch": 9.377292576419213,
+ "grad_norm": 0.026340370997786522,
+ "learning_rate": 0.002807222875684246,
+ "loss": 5.527161598205566,
+ "step": 675
+ },
+ {
+ "epoch": 9.391266375545852,
+ "grad_norm": 0.02200835384428501,
+ "learning_rate": 0.002806565582736208,
+ "loss": 5.507146835327148,
+ "step": 676
+ },
+ {
+ "epoch": 9.405240174672489,
+ "grad_norm": 0.020420534536242485,
+ "learning_rate": 0.0028059072491971085,
+ "loss": 5.5162506103515625,
+ "step": 677
+ },
+ {
+ "epoch": 9.419213973799126,
+ "grad_norm": 0.02493203990161419,
+ "learning_rate": 0.0028052478755973545,
+ "loss": 5.5252556800842285,
+ "step": 678
+ },
+ {
+ "epoch": 9.433187772925764,
+ "grad_norm": 0.021293973550200462,
+ "learning_rate": 0.0028045874624681925,
+ "loss": 5.523589134216309,
+ "step": 679
+ },
+ {
+ "epoch": 9.447161572052401,
+ "grad_norm": 0.016593236476182938,
+ "learning_rate": 0.0028039260103417045,
+ "loss": 5.500586032867432,
+ "step": 680
+ },
+ {
+ "epoch": 9.46113537117904,
+ "grad_norm": 0.018857723101973534,
+ "learning_rate": 0.002803263519750811,
+ "loss": 5.52933406829834,
+ "step": 681
+ },
+ {
+ "epoch": 9.475109170305677,
+ "grad_norm": 0.015930436551570892,
+ "learning_rate": 0.0028025999912292677,
+ "loss": 5.48813533782959,
+ "step": 682
+ },
+ {
+ "epoch": 9.489082969432314,
+ "grad_norm": 0.019427936524152756,
+ "learning_rate": 0.002801935425311669,
+ "loss": 5.46573543548584,
+ "step": 683
+ },
+ {
+ "epoch": 9.503056768558952,
+ "grad_norm": 0.018123053014278412,
+ "learning_rate": 0.0028012698225334427,
+ "loss": 5.499445915222168,
+ "step": 684
+ },
+ {
+ "epoch": 9.51703056768559,
+ "grad_norm": 0.02035410702228546,
+ "learning_rate": 0.002800603183430853,
+ "loss": 5.513540267944336,
+ "step": 685
+ },
+ {
+ "epoch": 9.531004366812226,
+ "grad_norm": 0.025838350877165794,
+ "learning_rate": 0.0027999355085409996,
+ "loss": 5.565637588500977,
+ "step": 686
+ },
+ {
+ "epoch": 9.544978165938865,
+ "grad_norm": 0.029621612280607224,
+ "learning_rate": 0.002799266798401816,
+ "loss": 5.5619215965271,
+ "step": 687
+ },
+ {
+ "epoch": 9.558951965065502,
+ "grad_norm": 0.028948379680514336,
+ "learning_rate": 0.0027985970535520685,
+ "loss": 5.551331520080566,
+ "step": 688
+ },
+ {
+ "epoch": 9.57292576419214,
+ "grad_norm": 0.022135930135846138,
+ "learning_rate": 0.0027979262745313604,
+ "loss": 5.527722358703613,
+ "step": 689
+ },
+ {
+ "epoch": 9.586899563318777,
+ "grad_norm": 0.028884069994091988,
+ "learning_rate": 0.0027972544618801256,
+ "loss": 5.554599761962891,
+ "step": 690
+ },
+ {
+ "epoch": 9.600873362445414,
+ "grad_norm": 0.02499174140393734,
+ "learning_rate": 0.00279658161613963,
+ "loss": 5.543351173400879,
+ "step": 691
+ },
+ {
+ "epoch": 9.614847161572053,
+ "grad_norm": 0.021131042391061783,
+ "learning_rate": 0.0027959077378519755,
+ "loss": 5.496039390563965,
+ "step": 692
+ },
+ {
+ "epoch": 9.62882096069869,
+ "grad_norm": 0.02957489714026451,
+ "learning_rate": 0.002795232827560092,
+ "loss": 5.552300453186035,
+ "step": 693
+ },
+ {
+ "epoch": 9.642794759825328,
+ "grad_norm": 0.022832119837403297,
+ "learning_rate": 0.0027945568858077433,
+ "loss": 5.488651275634766,
+ "step": 694
+ },
+ {
+ "epoch": 9.656768558951965,
+ "grad_norm": 0.02074701525270939,
+ "learning_rate": 0.002793879913139523,
+ "loss": 5.491244792938232,
+ "step": 695
+ },
+ {
+ "epoch": 9.670742358078602,
+ "grad_norm": 0.019919803366065025,
+ "learning_rate": 0.002793201910100856,
+ "loss": 5.483855247497559,
+ "step": 696
+ },
+ {
+ "epoch": 9.68471615720524,
+ "grad_norm": 0.01858169585466385,
+ "learning_rate": 0.002792522877237998,
+ "loss": 5.50787353515625,
+ "step": 697
+ },
+ {
+ "epoch": 9.698689956331878,
+ "grad_norm": 0.02085857465863228,
+ "learning_rate": 0.002791842815098032,
+ "loss": 5.512872695922852,
+ "step": 698
+ },
+ {
+ "epoch": 9.712663755458514,
+ "grad_norm": 0.01407450158149004,
+ "learning_rate": 0.002791161724228872,
+ "loss": 5.472590446472168,
+ "step": 699
+ },
+ {
+ "epoch": 9.726637554585153,
+ "grad_norm": 0.01672944612801075,
+ "learning_rate": 0.002790479605179262,
+ "loss": 5.469183921813965,
+ "step": 700
+ },
+ {
+ "epoch": 9.74061135371179,
+ "grad_norm": 0.016565700992941856,
+ "learning_rate": 0.0027897964584987713,
+ "loss": 5.493677139282227,
+ "step": 701
+ },
+ {
+ "epoch": 9.754585152838429,
+ "grad_norm": 0.017017360776662827,
+ "learning_rate": 0.0027891122847377996,
+ "loss": 5.464931488037109,
+ "step": 702
+ },
+ {
+ "epoch": 9.768558951965066,
+ "grad_norm": 0.019842486828565598,
+ "learning_rate": 0.002788427084447574,
+ "loss": 5.461293697357178,
+ "step": 703
+ },
+ {
+ "epoch": 9.782532751091702,
+ "grad_norm": 0.016210904344916344,
+ "learning_rate": 0.0027877408581801467,
+ "loss": 5.454716682434082,
+ "step": 704
+ },
+ {
+ "epoch": 9.796506550218341,
+ "grad_norm": 0.02064039371907711,
+ "learning_rate": 0.002787053606488398,
+ "loss": 5.466576099395752,
+ "step": 705
+ },
+ {
+ "epoch": 9.810480349344978,
+ "grad_norm": 0.018198711797595024,
+ "learning_rate": 0.0027863653299260356,
+ "loss": 5.447747230529785,
+ "step": 706
+ },
+ {
+ "epoch": 9.824454148471617,
+ "grad_norm": 0.015937626361846924,
+ "learning_rate": 0.00278567602904759,
+ "loss": 5.5055036544799805,
+ "step": 707
+ },
+ {
+ "epoch": 9.838427947598253,
+ "grad_norm": 0.01687259040772915,
+ "learning_rate": 0.002784985704408419,
+ "loss": 5.422250747680664,
+ "step": 708
+ },
+ {
+ "epoch": 9.85240174672489,
+ "grad_norm": 0.02622046135365963,
+ "learning_rate": 0.002784294356564705,
+ "loss": 5.466894626617432,
+ "step": 709
+ },
+ {
+ "epoch": 9.866375545851529,
+ "grad_norm": 0.018672186881303787,
+ "learning_rate": 0.0027836019860734535,
+ "loss": 5.48300838470459,
+ "step": 710
+ },
+ {
+ "epoch": 9.880349344978166,
+ "grad_norm": 0.02166241593658924,
+ "learning_rate": 0.0027829085934924958,
+ "loss": 5.477658271789551,
+ "step": 711
+ },
+ {
+ "epoch": 9.894323144104803,
+ "grad_norm": 0.018426023423671722,
+ "learning_rate": 0.0027822141793804856,
+ "loss": 5.487186431884766,
+ "step": 712
+ },
+ {
+ "epoch": 9.908296943231441,
+ "grad_norm": 0.022229891270399094,
+ "learning_rate": 0.0027815187442968995,
+ "loss": 5.488579273223877,
+ "step": 713
+ },
+ {
+ "epoch": 9.922270742358078,
+ "grad_norm": 0.026930972933769226,
+ "learning_rate": 0.002780822288802037,
+ "loss": 5.492000102996826,
+ "step": 714
+ },
+ {
+ "epoch": 9.936244541484717,
+ "grad_norm": 0.02733265422284603,
+ "learning_rate": 0.00278012481345702,
+ "loss": 5.4663801193237305,
+ "step": 715
+ },
+ {
+ "epoch": 9.950218340611354,
+ "grad_norm": 0.02993253618478775,
+ "learning_rate": 0.0027794263188237922,
+ "loss": 5.543368816375732,
+ "step": 716
+ },
+ {
+ "epoch": 9.96419213973799,
+ "grad_norm": 0.03152153640985489,
+ "learning_rate": 0.002778726805465118,
+ "loss": 5.565644264221191,
+ "step": 717
+ },
+ {
+ "epoch": 9.97816593886463,
+ "grad_norm": 0.031395044177770615,
+ "learning_rate": 0.002778026273944581,
+ "loss": 5.553359508514404,
+ "step": 718
+ },
+ {
+ "epoch": 9.992139737991266,
+ "grad_norm": 0.026603225618600845,
+ "learning_rate": 0.002777324724826589,
+ "loss": 5.552347183227539,
+ "step": 719
+ },
+ {
+ "epoch": 10.0,
+ "grad_norm": 0.024152139201760292,
+ "learning_rate": 0.002776622158676366,
+ "loss": 5.493257522583008,
+ "step": 720
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 3600,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 50,
+ "save_steps": 500,
+ "stateful_callbacks": {
+ "TrainerControl": {
+ "args": {
+ "should_epoch_stop": false,
+ "should_evaluate": true,
+ "should_log": false,
+ "should_save": false,
+ "should_training_stop": false
+ },
+ "attributes": {}
+ }
+ },
+ "total_flos": 3.069613361332224e+17,
+ "train_batch_size": 16,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/runs/baseline/checkpoint-720/training_args.bin b/runs/baseline/checkpoint-720/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7
--- /dev/null
+++ b/runs/baseline/checkpoint-720/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97
+size 5265